| [0.000 --> 22.200] Okay, now I don't want to alarm anybody in this room, but it's just come to my attention | |
| [22.200 --> 26.600] that the person to your right is a liar. | |
| [26.600 --> 30.440] So the person to your left is a liar. | |
| [30.440 --> 33.640] Also the person sitting in your very seats is a liar. | |
| [33.640 --> 35.440] We're all liars. | |
| [35.440 --> 39.080] What I'm going to do today is I'm going to show you what the research says about why we're | |
| [39.080 --> 44.320] all liars, how you can become a lie spotter, and why you might want to go the extra mile | |
| [44.320 --> 47.720] and go from lie spotting to truth seeking. | |
| [47.720 --> 50.520] And ultimately to trust building. | |
| [50.520 --> 56.560] Now speaking of trust, ever since I wrote this book lie spotting, no one wants to meet | |
| [56.560 --> 57.560] me in person anymore. | |
| [57.560 --> 58.560] No, no, no, no. | |
| [58.560 --> 60.560] They say it's okay. | |
| [60.560 --> 63.320] We'll email you. | |
| [63.320 --> 67.680] I can't even get a coffee date at Starbucks. | |
| [67.680 --> 69.680] My husband's like honey deception. | |
| [69.680 --> 71.280] Maybe you could have focused on cooking. | |
| [71.280 --> 72.960] How about French cooking? | |
| [72.960 --> 78.200] So before we get started, what I'm going to do is I'm going to clarify my goal for you, | |
| [78.200 --> 80.320] which is not to teach a game of gacha. | |
| [80.320 --> 83.600] Lie spotters aren't those nit picky kids, those kids in the back of the room that are | |
| [83.600 --> 88.600] shouting gacha, gacha, your eyebrow twitched, you flared, you're an astral, I watched that | |
| [88.600 --> 91.200] TV show lie to me, I know you're lying. | |
| [91.200 --> 92.200] No. | |
| [92.200 --> 96.360] Lie spotters are armed with scientific knowledge of how to spot deception. | |
| [96.360 --> 100.560] They use it to get to the truth, and they do what mature leaders do every day. | |
| [100.560 --> 105.600] They have difficult conversations with difficult people, sometimes during very difficult times. | |
| [105.600 --> 109.200] And they start up that path by accepting a core proposition. | |
| [109.200 --> 111.440] And that proposition is the filing. | |
| [111.440 --> 114.600] Lying is a cooperative act. | |
| [114.600 --> 115.600] Think about it. | |
| [115.600 --> 118.120] A lie has no power whatsoever by its mere utterance. | |
| [118.120 --> 121.960] Its power emerges when someone else agrees to believe the lie. | |
| [121.960 --> 128.320] So I know it may sound like tough love, but look, if at some point you got lied to, it's | |
| [128.320 --> 130.200] because you agreed to get lied to. | |
| [130.200 --> 131.720] Truth number one about lying. | |
| [131.720 --> 132.760] Lying is a cooperative act. | |
| [132.760 --> 135.840] Now, not all lies are harmful. | |
| [135.840 --> 141.360] Sometimes we're willing participants in deception for the sake of social dignity, | |
| [141.360 --> 144.160] maybe to keep a secret that should be kept secret secret. | |
| [144.160 --> 146.160] We say, nice song. | |
| [146.160 --> 147.760] Honey, you don't look fat in that. | |
| [147.760 --> 148.760] No. | |
| [148.760 --> 151.200] Or we say, favorite of the digerati. | |
| [151.200 --> 154.280] You know, I just fished that email out of my spam folder. | |
| [154.280 --> 156.720] I'm so sorry. | |
| [156.720 --> 160.360] But there are times when we are unwilling participants in deception. | |
| [160.360 --> 162.800] And that can have dramatic costs for us. | |
| [162.800 --> 170.160] Last year saw $997 billion in corporate fraud alone in the United States. | |
| [170.160 --> 172.200] That's an eyelash under $1 trillion. | |
| [172.200 --> 174.480] That's 7% of revenues. | |
| [174.480 --> 175.560] Deception can cost billions. | |
| [175.560 --> 181.560] Think Enron, Mat off the mortgage crisis, or in the case of double agents and traders | |
| [181.560 --> 185.960] like Robert Hansen or Aldrich Ames, lies can betray our country. | |
| [185.960 --> 188.080] They can compromise our security. | |
| [188.080 --> 189.360] They can undermine democracy. | |
| [189.360 --> 191.960] They can cause the deaths of those that defend us. | |
| [191.960 --> 194.720] Deception is actually serious business. | |
| [194.720 --> 199.640] This con man, Henry Oberlander, he was such an effective con man. | |
| [199.640 --> 203.360] The British authorities say he could have undermined the entire banking system of the Western | |
| [203.360 --> 204.360] world. | |
| [204.360 --> 205.360] And you can't find this guy on Google. | |
| [205.360 --> 206.360] You can't find him anywhere. | |
| [206.360 --> 207.360] He was interviewed once. | |
| [207.360 --> 208.360] And he said the following. | |
| [208.360 --> 210.520] He said, look, I've got one rule. | |
| [210.520 --> 211.680] And this was Henry's rule. | |
| [211.680 --> 215.600] He said, look, everyone is willing to give you something. | |
| [215.600 --> 219.480] They're ready to give you something for whatever it is they're hungry for. | |
| [219.480 --> 220.480] And that's the crux of it. | |
| [220.480 --> 225.360] If you don't want to be deceived, you have to know what is it that you're hungry for. | |
| [225.360 --> 227.120] And we all kind of hate to admit it. | |
| [227.120 --> 233.000] You know, we kind of wish we were better husbands, better wives, smarter, more powerful, | |
| [233.000 --> 234.560] taller, richer. | |
| [234.560 --> 237.000] The list goes on. | |
| [237.000 --> 241.640] Lying is an attempt to bridge that gap, to connect our wishes and our fantasies about | |
| [241.640 --> 245.720] who we wish we were, how we wish we could be, with what we're really like. | |
| [245.720 --> 249.200] And, boy, we're willing to fill in those gaps in our lives with lies. | |
| [249.200 --> 254.680] On a given day, studies show that you may be lied to anywhere from 10 to 200 times. | |
| [254.680 --> 257.960] And now granted, many of those are white lies. | |
| [257.960 --> 263.600] But in another study, it showed that strangers lied three times within the first 10 minutes | |
| [263.600 --> 265.040] of meeting each other. | |
| [265.040 --> 269.320] Now, when we first hear this data, we recoil. | |
| [269.320 --> 270.760] We can't believe how prevalent lying is. | |
| [270.760 --> 272.560] We're essentially against lying. | |
| [272.560 --> 276.720] But if you look more closely, the plot actually thickens. | |
| [276.720 --> 280.640] We lie more to strangers than we lie to co-workers. | |
| [280.640 --> 284.400] Extroverts lie more than introverts. | |
| [284.400 --> 289.480] And lie eight times more about themselves than they do other people. | |
| [289.480 --> 292.600] Women lie more to protect other people. | |
| [292.600 --> 297.200] If you're in an average married couple, you're going to lie to your spouse in one out of | |
| [297.200 --> 298.600] every 10 interactions. | |
| [298.600 --> 300.400] Now, you may think that's bad. | |
| [300.400 --> 303.760] If you're unmarried, that number drops to three. | |
| [303.760 --> 305.240] Lying's complex. | |
| [305.240 --> 308.880] It's woven into the fabric of our daily and our business lives, where deeply ambivalent | |
| [308.880 --> 310.120] about the truth. | |
| [310.120 --> 314.240] We parse it out on an as-needed basis, sometimes for very, very good reasons, and other times | |
| [314.640 --> 317.720] just because we don't understand the gaps in our lives. | |
| [317.720 --> 319.360] That's truth number two about lying. | |
| [319.360 --> 322.680] We're against lying, but we're covertly for it. | |
| [322.680 --> 326.720] It weighs that our society has sanctioned for centuries and centuries and centuries. | |
| [326.720 --> 328.680] It's as old as breathing. | |
| [328.680 --> 329.680] It's part of our culture. | |
| [329.680 --> 331.400] It's part of our history. | |
| [331.400 --> 338.480] Think Dante, Shakespeare, the Bible, news of the world. | |
| [338.480 --> 341.480] Lying has evolutionary value to us as a species. | |
| [341.480 --> 346.520] Researchers have long known that the more intelligent the species, the larger the neocortex, | |
| [346.520 --> 349.200] the more likely it is to be deceptive. | |
| [349.200 --> 350.680] Now you might remember Coco. | |
| [350.680 --> 354.000] Does anybody here remember Coco the Gorilla, who was taught sign language? | |
| [354.000 --> 357.200] Coco was taught to communicate via sign language. | |
| [357.200 --> 358.200] Here's Coco with her kitten. | |
| [358.200 --> 361.800] It's her cute little fluffy pet kitten. | |
| [361.800 --> 368.000] Coco once blamed her pet kitten for ripping a sink out of the wall. | |
| [368.000 --> 370.200] We're hardwired to become leaders of the pack. | |
| [370.200 --> 372.240] It starts really, really early. | |
| [372.240 --> 373.240] How early? | |
| [373.240 --> 378.480] Well, babies will fake a cry, pause, wait to see who's coming, and then go right back to | |
| [378.480 --> 380.560] crying. | |
| [380.560 --> 383.760] One year old's learned concealment. | |
| [383.760 --> 386.400] Two year old's bluff. | |
| [386.400 --> 387.760] Five year olds lie outright. | |
| [387.760 --> 390.200] They manipulate via flattery. | |
| [390.200 --> 392.720] Nine year olds masters of the cover up. | |
| [392.720 --> 395.680] By the time you enter college, you're going to lie to your mom and one out of every five | |
| [395.680 --> 397.880] interactions. | |
| [397.880 --> 401.880] By the time we enter this work world and we're breadwinners, we enter a world that has | |
| [401.880 --> 408.960] just cluttered with spam, fake digital friends, partisan media, ingenious identity thieves, | |
| [408.960 --> 413.120] world-class Ponzi schemers, a deception epidemic. | |
| [413.120 --> 417.960] In short, what one author calls a post-truth society. | |
| [417.960 --> 424.200] It's been very confusing for a long time now. | |
| [424.200 --> 426.440] What do you do? | |
| [426.440 --> 430.600] Well, there are steps we can take to navigate our way through the morass. | |
| [430.600 --> 433.120] Trained lie spotters get to the truth 90% of the time. | |
| [433.120 --> 435.920] The rest of us are only 54% accurate. | |
| [435.920 --> 437.520] Why is it so easy to learn? | |
| [437.520 --> 439.520] Well, they're good liars and they're bad liars. | |
| [439.520 --> 440.920] They're no real original liars. | |
| [440.920 --> 442.520] We all bake the same mistakes. | |
| [442.520 --> 444.520] We all use the same techniques. | |
| [444.520 --> 448.400] So what I'm going to do is I'm going to show you two patterns of deception and then we're | |
| [448.400 --> 450.600] going to look at the hot spots and see if we can find them ourselves. | |
| [450.600 --> 454.600] We're going to start with speech. | |
| [454.600 --> 455.760] I want you to listen to me. | |
| [455.760 --> 458.000] I'm going to say this again. | |
| [458.000 --> 464.520] I did not have sexual relations with that woman, Miss Lewinsky. | |
| [464.520 --> 469.680] I never told anybody it's a lie, not a single time, never. | |
| [469.680 --> 475.160] These allegations are false and I need to go back to work for the American people. | |
| [475.160 --> 477.160] Thank you. | |
| [477.160 --> 479.880] Okay. | |
| [479.880 --> 481.400] What were the telltale signs? | |
| [481.400 --> 486.280] Well, first we heard what's known as a non-contracted denial. | |
| [486.280 --> 490.920] Studies show that people who are over-determined in their denial will resort to formal rather than | |
| [490.920 --> 492.360] informal language. | |
| [492.360 --> 495.000] We also heard distancing language that woman. | |
| [495.000 --> 499.920] We know that liars will unconsciously distance themselves from their subject using languages | |
| [499.920 --> 500.920] their tool. | |
| [500.920 --> 505.080] Now, Bill Clinton has said, well, they tell you the truth. | |
| [505.080 --> 507.480] Or Richard Nixon's favorite in all candor. | |
| [507.480 --> 511.160] He would have been a dead giveaway for any lie spotter that knows the qualifying language | |
| [511.240 --> 513.560] as it's called qualifying language like that. | |
| [513.560 --> 516.240] Further discredits the subject. | |
| [516.240 --> 521.000] Now if he had repeated the question in its entirety or if he had peppered his account with a little | |
| [521.000 --> 525.880] too much detail and we're all really glad he didn't do that, he would have further discredited | |
| [525.880 --> 527.640] himself. | |
| [527.640 --> 528.640] Freud had it right. | |
| [528.640 --> 532.520] Freud said, look, there's much more to it than speech. | |
| [532.520 --> 534.920] No mortal can keep a secret. | |
| [534.920 --> 538.960] If his lips are silent, he shatters with his fingertips and we all do it. | |
| [538.960 --> 541.120] No matter how powerful you are, we all do it. | |
| [541.120 --> 543.000] We all chatter with our fingertips. | |
| [543.000 --> 550.400] I'm going to show you Dominique Strauss-Con with Obama, who's chattering with his fingertips. | |
| [550.400 --> 558.280] Now this brings us to our next pattern, which is body language. | |
| [558.280 --> 560.640] With body language, here's what you've got to do. | |
| [560.640 --> 565.120] You really got to just throw your assumptions out the door, let the science temper your | |
| [565.120 --> 568.840] knowledge a little bit because we think liars fidget all the time. | |
| [568.840 --> 572.240] Well guess what, they're known to freeze their upper bodies when they're lying. | |
| [572.240 --> 575.040] We think liars won't look in the eyes. | |
| [575.040 --> 578.600] Well guess what, they look you in the eyes a little too much just to compensate for that | |
| [578.600 --> 579.600] myth. | |
| [579.600 --> 585.760] We think warmth and smiles convey honesty, sincerity, but a trained lie spotter can spot a fake | |
| [585.760 --> 587.400] smile a mile away. | |
| [587.400 --> 591.160] Can you all spot the fake smile here? | |
| [591.160 --> 596.960] You can consciously contract the muscles in your cheeks, but the real smiles in the | |
| [596.960 --> 597.960] eyes. | |
| [597.960 --> 602.460] Those feet of the eyes, they cannot be consciously contracted, especially if you overdo the | |
| [602.460 --> 603.460] Botox. | |
| [603.460 --> 604.460] Don't overdo the Botox. | |
| [604.460 --> 606.760] Nobody will think you're honest. | |
| [606.760 --> 608.600] Now we're going to look at the hot spots. | |
| [608.600 --> 610.400] Can you tell what's happening in a conversation? | |
| [610.400 --> 615.280] Can you start to find the hot spots to see the discrepancies between someone's words and | |
| [615.280 --> 616.440] someone's actions? | |
| [616.440 --> 620.880] Now I know it seems really obvious, but when you're having a conversation with someone | |
| [620.880 --> 625.920] that uses a speck of deception, attitude is by far the most overlooked, but telling of | |
| [625.920 --> 627.160] indicators. | |
| [627.160 --> 631.040] When honest person is going to be cooperative, they're going to show their own side, they're | |
| [631.040 --> 634.200] going to be enthusiastic, they're going to be willing and helpful in getting you to | |
| [634.200 --> 635.200] the truth. | |
| [635.200 --> 639.480] They're going to be willing to brainstorm, name suspects, provide details. | |
| [639.480 --> 644.920] They're going to say, hey, maybe it was those guys in payroll that forged those checks. | |
| [644.920 --> 648.600] They're going to be infuriated if they sense the wrongly accused about the entire course | |
| [648.600 --> 652.040] of the interview, not just in flashes, they'll be infuriated throughout the entire course | |
| [652.040 --> 653.040] of the interview. | |
| [653.560 --> 658.200] If you ask someone honest, what should happen to whoever did forge those checks? | |
| [658.200 --> 663.280] An honest person is much more likely to recommend strict, rather than lenient punishment. | |
| [663.280 --> 668.440] Now, let's say you're having that exact same conversation with someone deceptive. | |
| [668.440 --> 674.680] That person may be withdrawn, look down, lower their voice, pause, be kind of herky jerky. | |
| [674.680 --> 678.000] Ask a deceptive person to tell their story, they're going to pepper it with way too much | |
| [678.000 --> 681.840] detail in all kinds of irrelevant places. | |
| [681.840 --> 684.640] And then they're going to tell their story in strict, chronological order. | |
| [684.640 --> 689.520] And what a trained interrogator does is they come in, and in very subtle ways, over the | |
| [689.520 --> 694.280] course of several hours, they will ask that person to tell their story backwards. | |
| [694.280 --> 698.600] And then they'll watch them squirm and track which questions produce the highest volume | |
| [698.600 --> 699.600] of deceptive tells. | |
| [699.600 --> 700.600] Why do they do that? | |
| [700.600 --> 702.360] Well, we all do the same thing. | |
| [702.360 --> 705.720] We rehearse our words, but we rarely rehearse our gestures. | |
| [705.720 --> 707.920] We say, yes, we shake our heads, no. | |
| [707.920 --> 709.280] We tell very convincing stories. | |
| [709.280 --> 711.160] We slightly shrug our shoulders. | |
| [711.160 --> 715.840] We commit terrible crimes, and we smile at the delight in getting away with it. | |
| [715.840 --> 718.920] Now that smile is known in the trade as duping delight. | |
| [718.920 --> 722.680] And we're going to see that in several videos moving forward, but we're going to start | |
| [722.680 --> 724.080] for those of you that don't know him. | |
| [724.080 --> 728.840] This is presidential candidate John Edwards, who shocked America by fathering a child out | |
| [728.840 --> 729.840] of wedlock. | |
| [729.840 --> 733.080] We're going to see him talk about getting a paternity test. | |
| [733.080 --> 738.080] To see now, if you can spot him saying yes while shaking his head, no, slightly shrugging | |
| [738.080 --> 739.080] his shoulders. | |
| [739.080 --> 741.320] I'm happy to participate in one. | |
| [741.320 --> 746.760] I know that it's not possible that this child could be mine because of the timing of events. | |
| [746.760 --> 748.480] So I know it's not possible. | |
| [748.480 --> 751.520] Happy to take a paternity test, and we'd love to see it happen. | |
| [751.520 --> 752.920] Are you going to do that soon? | |
| [752.920 --> 754.920] Is there somebody that you can't? | |
| [754.920 --> 755.920] I'm only one side. | |
| [755.920 --> 760.760] I can only one side of the test, but I'm happy to participate in one. | |
| [760.760 --> 765.640] Okay, those head shakes are much easier to spot once you know to look for them. | |
| [765.640 --> 770.320] Now there are going to be times when someone makes one expression while masking another that | |
| [770.320 --> 773.000] just kind of leaks through in a flash. | |
| [773.000 --> 775.680] Murderers are known to leak sadness. | |
| [775.680 --> 778.640] Your new joint venture partner might shake your hand, celebrate, go out to dinner with | |
| [778.640 --> 781.720] you, and then leak an expression of anger. | |
| [781.720 --> 785.400] And we're not all going to become facial expression experts overnight here, but there's one | |
| [785.400 --> 789.120] I can teach you that's very dangerous and that's easy to learn, and that's the expression | |
| [789.120 --> 790.120] of contempt. | |
| [790.120 --> 794.160] Now with anger, you've got two people in an even playing field. | |
| [794.160 --> 798.480] It's still somewhat of a healthy relationship, but when anger turns to contempt, you've | |
| [798.480 --> 800.080] been dismissed. | |
| [800.080 --> 804.800] It's associated with moral superiority, and for that reason, it's very, very hard to | |
| [804.800 --> 805.800] recover from. | |
| [805.800 --> 807.320] Here's what it looks like. | |
| [807.320 --> 811.360] It's marked by one lip corner pulled up and in. | |
| [811.360 --> 814.200] It's the only asymmetrical expression. | |
| [814.200 --> 818.640] And in the presence of contempt, whether or not deception follows, and it doesn't always | |
| [818.640 --> 820.120] follow. | |
| [820.120 --> 824.080] Look the other way, go the other direction, reconsider the deal, say no, thank you. | |
| [824.080 --> 826.680] I'm not coming up for just one more night cap. | |
| [826.680 --> 828.320] Thank you. | |
| [828.320 --> 831.840] Science has surfaced many, many more indicators. | |
| [831.840 --> 836.680] We know, for example, we know Lyra's will shift their blink rate, point their feet towards | |
| [836.680 --> 838.440] an exit. | |
| [838.440 --> 842.760] They will take barrier objects and put them between themselves and the person that's interviewing | |
| [842.760 --> 843.760] them. | |
| [843.760 --> 847.680] They'll alter their vocal tone, often making them make their vocal tone much lower. | |
| [847.680 --> 850.080] Now, here's the deal. | |
| [850.080 --> 853.480] These behaviors are just behaviors. | |
| [853.480 --> 856.040] They're not proof of deception. | |
| [856.040 --> 857.040] They're red flags. | |
| [857.040 --> 858.040] We're human beings. | |
| [858.040 --> 861.520] We make deceptive flailing gestures all over the place all day long. | |
| [861.520 --> 863.920] They don't mean anything in and of themselves. | |
| [863.920 --> 867.000] But when you see clusters of them, that's your signal. | |
| [867.000 --> 868.760] Look, listen, probe. | |
| [868.760 --> 870.080] Ask some hard questions. | |
| [870.080 --> 872.960] Get out of that very comfortable mode of knowing. | |
| [872.960 --> 874.600] Walk into curiosity mode. | |
| [874.600 --> 876.480] Ask more questions. | |
| [876.480 --> 877.480] Have a little dignity. | |
| [877.480 --> 878.960] Treat the person you're talking to with rapport. | |
| [878.960 --> 882.320] Don't try to be like those folks on law and order and those other TV shows that pummel | |
| [882.320 --> 884.080] their subjects into submission. | |
| [884.080 --> 885.080] Don't be too aggressive. | |
| [885.080 --> 887.040] It doesn't work. | |
| [887.040 --> 890.720] Now, we've talked a little bit about how to talk to someone who's lying. | |
| [890.720 --> 892.040] I mean, how to spot a lie. | |
| [892.040 --> 895.160] And as I promise, we're now going to look at what the truth looks like. | |
| [895.160 --> 897.920] And I'm going to show you two videos. | |
| [897.920 --> 899.280] Two mothers. | |
| [899.280 --> 900.280] One is lying. | |
| [900.280 --> 901.280] One is telling the truth. | |
| [901.280 --> 904.600] And these were surfaced by researcher David Matsumoto in California. | |
| [904.600 --> 908.680] And I think they're an excellent example of what the truth looks like. | |
| [908.680 --> 913.640] This mother, Diane Downes, shot her kids at close range. | |
| [913.640 --> 917.040] Draw them to the hospital while they bled all over the car. | |
| [917.040 --> 919.520] Claimed a scraggie-haired stranger did it. | |
| [919.520 --> 923.280] And you'll see when you see the video, she can't even pretend to be an agonizing mother. | |
| [923.280 --> 928.340] What you want to look for here is an incredible discrepancy between horrific events that she | |
| [928.340 --> 931.400] describes and her very, very cool demeanor. | |
| [931.400 --> 934.320] And if you look closely, you'll see dooping delight throughout this video. | |
| [934.320 --> 937.880] But at night, when I close my eyes, I can see Christy reaching her hand out to me while | |
| [937.880 --> 939.480] I'm driving. | |
| [939.480 --> 942.000] And the blood just keeps coming out of her mouth. | |
| [942.000 --> 944.720] And that, maybe it'll fade too with time. | |
| [944.720 --> 946.440] But I don't think so. | |
| [946.440 --> 956.000] That haunts me the most. | |
| [956.000 --> 960.680] Now I'm going to show you a video of an actual grieving mother, Erin Runyon, confronting | |
| [960.680 --> 963.720] her daughter's murder and torture in court. | |
| [963.720 --> 965.680] Here you're going to see no false emotion. | |
| [965.680 --> 968.880] The authentic expression of a mother's agony. | |
| [968.880 --> 973.640] I wrote the statement on the third anniversary of the night you took my baby. | |
| [973.640 --> 974.640] And you heard her. | |
| [974.640 --> 976.640] Then you crushed her. | |
| [976.640 --> 978.640] You terrified her. | |
| [978.640 --> 981.120] And tell her her heart stopped. | |
| [981.120 --> 982.640] And she fought. | |
| [982.640 --> 984.640] And I know she fought you. | |
| [984.640 --> 987.840] But I know she looked at you with those amazing brown eyes. | |
| [987.840 --> 991.600] And you still wanted to kill her. | |
| [991.600 --> 993.040] And I don't understand it. | |
| [993.040 --> 994.440] And they never will. | |
| [994.440 --> 1000.000] OK, there's no doubting the veracity of those emotions. | |
| [1000.000 --> 1005.920] Now the technology around what the truth looks like is progressing on the science of it. | |
| [1005.920 --> 1012.120] We know, for example, that we now have specialized eye trackers, infrared brain scans, MRIs that | |
| [1012.120 --> 1016.360] can decode the signals that our bodies send out when we're trying to be deceptive. | |
| [1016.360 --> 1020.960] And these technologies are going to be marketed to all of us as panacea is for deceit. | |
| [1020.960 --> 1022.960] And they will prove incredibly useful someday. | |
| [1022.960 --> 1028.080] But you've got to ask yourself in the meantime, who do you want on your side of the meeting? | |
| [1028.080 --> 1032.280] Someone who's trained in getting to the truth or some guy is going to drag a 400 pound | |
| [1032.280 --> 1035.520] electosephalogram through the door. | |
| [1035.520 --> 1039.560] Lies spotters rely on human tools. | |
| [1039.560 --> 1043.160] They know, as someone once said, characters who you are in the dark. | |
| [1043.160 --> 1047.920] And what's kind of interesting is that today we have so little darkness. | |
| [1047.920 --> 1050.160] Our world has lit up 24 hours a day. | |
| [1050.160 --> 1052.000] It's transparent. | |
| [1052.000 --> 1056.240] With blogs and social networks broadcasting the buzz of a whole new generation of people | |
| [1056.240 --> 1059.240] that have made a choice to live their lives in public. | |
| [1059.240 --> 1063.520] It's a much more noisy world. | |
| [1063.520 --> 1070.360] So one challenge we have is to remember, oversharing, that's not honesty. | |
| [1070.360 --> 1075.360] Our manic tweeting and texting can blind us to the fact that the subtleties of human | |
| [1075.360 --> 1078.120] decency, character, integrity, that's still what matters. | |
| [1078.120 --> 1079.960] That's always what's going to matter. | |
| [1079.960 --> 1086.160] So in this much noisy world, it might make sense for us to be just a little bit more explicit | |
| [1086.160 --> 1088.840] about our moral code. | |
| [1088.840 --> 1093.160] When you combine the science of recognizing deception with the art of looking, listening, | |
| [1093.160 --> 1096.600] you exempt yourself from collaborating in a lie. | |
| [1096.600 --> 1101.240] You start up that path of being just a little bit more explicit because you signal to everyone | |
| [1101.240 --> 1102.240] around you. | |
| [1102.240 --> 1105.280] You say, hey, my world, our world. | |
| [1105.280 --> 1106.720] It's going to be an honest one. | |
| [1106.720 --> 1110.720] My world is going to be one where truth is strengthened and falsehood is recognized and | |
| [1110.720 --> 1112.200] marginalized. | |
| [1112.200 --> 1117.560] And when you do that, the ground around you starts to shift just a little bit. | |
| [1117.560 --> 1118.560] And that's the truth. | |
| [1118.560 --> 1119.560] Thank you. | |