April 30, 2020

#CX 50: The Marketing Impacts of Artificial Intelligence w/ William Ammerman

In this episode of the #CX series, Ethan Beute, Chief Evangelist at BombBomb, talks with William Ammerman, author of The Invisible Brand, about the impacts of artificial intelligence and machine learning on you as both a consumer and an operator inside a business.

Listen to more conversations on Ethan's podcast, The Customer Experience Podcast by clicking through one of the following links to your favorite podcast player:

Apple Podcasts

Spotify

Sticher

Google Play

Google Podcasts

Transcript
WEBVTT 1 00:00:04.679 --> 00:00:11.910 It's about persuasion, not coercion or deception. It's a simple formula. Most 2 00:00:12.070 --> 00:00:18.910 consumers are willing to be persuaded. As we move into a time when our 3 00:00:18.949 --> 00:00:25.500 business relationships are increasingly driven by and influenced by artificial intelligence and machine learning, 4 00:00:25.780 --> 00:00:31.539 it's important to understand the consequences. It's important to deploy tech in a way 5 00:00:31.579 --> 00:00:36.090 that puts our people in their best position to win. Mightime is Ethan butte 6 00:00:36.450 --> 00:00:42.090 coauthor of the Book Rehumanize Your Business and I also host the CX series here 7 00:00:42.170 --> 00:00:46.729 on b Tob Growth and our guest, William Ammerman, wrote an excellent book 8 00:00:47.170 --> 00:00:52.119 on the ethics and impacts of artificial intelligence. He breaks it down into common 9 00:00:52.159 --> 00:00:57.039 language and he'll get you thinking about what ai means for you as a customer 10 00:00:57.479 --> 00:01:03.000 and as an operator inside a business. Here we go, all right, 11 00:01:03.000 --> 00:01:07.150 I'm opening this episode of the Customer Experience Podcast with a quote, and here 12 00:01:07.189 --> 00:01:11.950 it is. As our interactions with artificially intelligent agents become more human like through 13 00:01:11.989 --> 00:01:18.430 natural language Algorithms, we will begin to have conversations with machines. We will 14 00:01:18.469 --> 00:01:23.540 build empathic relationships with those machines in which we are even more vulnerable to their 15 00:01:23.700 --> 00:01:26.620 influence. Those words come from a book that you should read. It's called 16 00:01:26.659 --> 00:01:32.060 the invisible and marketing in the age of automation, big data and machine learning 17 00:01:32.140 --> 00:01:36.810 and obviously has implications for the entire customer life cycle, marketing, sales and 18 00:01:36.890 --> 00:01:41.329 customer success. But it's author, and our guest today, also covers its 19 00:01:41.370 --> 00:01:46.689 implications for finance and investing, policy and politics, health and medicine, insurance, 20 00:01:46.730 --> 00:01:49.879 education, religion, even sex. It's a book I highly recommend and 21 00:01:49.920 --> 00:01:55.200 it draws on his master's work at UNC Chapel Hill about persuasion as a science 22 00:01:55.480 --> 00:01:59.640 and machine learnings ability to keep our faces glued to our screens. It also 23 00:01:59.680 --> 00:02:04.510 draws on his work at MIT on natural language processing and humans and pathic responses 24 00:02:04.670 --> 00:02:08.229 to our devices. He has years of experience as a VP and EEDP and 25 00:02:08.270 --> 00:02:14.389 digital media and digital advertising. He's currently the Executive Vice President of digital media 26 00:02:14.430 --> 00:02:19.699 at the engaged media portfolio of brands. William Ammerman, welcome to the customer 27 00:02:19.740 --> 00:02:24.340 experience podcast here. I'm glad to have the opportunity to talk about the invisible 28 00:02:24.379 --> 00:02:28.340 brand with you. As I said, I really, really enjoyed it. 29 00:02:28.379 --> 00:02:31.330 I recommend it to anyone who's listening there's a great deal of tension in there, 30 00:02:31.610 --> 00:02:36.530 and by the tension I mean, you know, the human experience machines. 31 00:02:36.569 --> 00:02:39.770 As they get smarter and they know more about us, they can leverage 32 00:02:39.810 --> 00:02:45.840 that against us, potentially in this vulnerability of this empathic response that we have 33 00:02:45.960 --> 00:02:49.120 an all these things just so rich. So, before we get going, 34 00:02:49.319 --> 00:02:55.039 let's cut right to it. is human domination by SUPERINTELLIGENCE INEVITABLE? No, 35 00:02:55.240 --> 00:03:00.830 I don't think so. I think a partnership between humans and machines is probably 36 00:03:00.870 --> 00:03:05.870 a better way of expressing what I view as the future. But I do 37 00:03:06.030 --> 00:03:09.430 think that machines get smarter and smarter and I think that machines will continue to 38 00:03:09.469 --> 00:03:15.699 do things that we associate with human like intelligence and that we are just really 39 00:03:16.020 --> 00:03:20.699 starting to experience it, so that we've just kind of seen the tip of 40 00:03:20.740 --> 00:03:23.740 the iceberg. What are your thoughts? When I see customer experiencing? You're 41 00:03:23.780 --> 00:03:28.889 preparing to come on to the customer experience podcast. What does customer experience mean 42 00:03:28.969 --> 00:03:31.409 to you? I don't have a pat definition, but I think what I 43 00:03:31.490 --> 00:03:37.810 think of the customer experience. I think of the interaction between brands and consumers 44 00:03:38.289 --> 00:03:44.360 and I think about whether or not the consumer has a positive or negative impression 45 00:03:44.400 --> 00:03:50.439 of the brand and I also think about whether or not the relationship feels ethical 46 00:03:50.639 --> 00:03:53.840 and transparent and trusted. So when I think of, you know, kind 47 00:03:53.879 --> 00:03:57.949 of the customer decision journey, which is a phrase that we use a lot 48 00:03:58.069 --> 00:04:02.310 in marketing, I think the customer experience is kind of informing the customer decision 49 00:04:02.349 --> 00:04:08.949 journey throughout the relationship between the brand and the and the buyer, the consumer. 50 00:04:09.590 --> 00:04:13.099 I like that you tapped there one of those kind of tension points that 51 00:04:13.180 --> 00:04:16.540 I felt in reading the invisible brand, which is this the transparency element, 52 00:04:16.699 --> 00:04:19.899 right, like what's inside the black box? Why am I getting these particular 53 00:04:19.980 --> 00:04:24.370 recommendations? Why am I seeing these particular ads and all of that, and 54 00:04:24.490 --> 00:04:29.009 we will get into that. There's one explicit reference to customer experience in the 55 00:04:29.089 --> 00:04:32.569 book, in those words, and it was around a negative experience with a 56 00:04:32.610 --> 00:04:35.930 fifty percent off sale. Do you want to tell that story just quickly? 57 00:04:38.250 --> 00:04:42.639 Oh boy, yeah, so my family was visiting me. I had an 58 00:04:42.759 --> 00:04:46.199 apartment up in New York City and around Christmas time my family came up and 59 00:04:46.279 --> 00:04:51.519 we did some shopping in the city together and my daughter spied a coat in 60 00:04:51.639 --> 00:04:58.829 the window of a Columbia store and in the meat packing district, I guess 61 00:04:58.910 --> 00:05:03.550 we were, and the kind of near Chelsea market. Anyway, the window, 62 00:05:03.990 --> 00:05:09.220 you know, clearly showed a fifty percent off sign with a rack of 63 00:05:09.339 --> 00:05:12.620 coads. We went in, my daughter tried one on, you know, 64 00:05:12.779 --> 00:05:15.939 retail price on it was like three hundred and fifty dollars, which was more 65 00:05:15.980 --> 00:05:19.500 than I was planning on spending on a single present. And but she said, 66 00:05:19.540 --> 00:05:23.370 she pointed out it was fifty percent off and I thought was still pretty 67 00:05:23.370 --> 00:05:26.170 expensive, but you know, she needs a winner code. And okay. 68 00:05:26.569 --> 00:05:30.810 So we took it up to the cash register and they rang it up and 69 00:05:30.730 --> 00:05:34.689 it was not fifty percent off and I said it's on the rack that says 70 00:05:34.720 --> 00:05:39.240 fifty percent off and and the guy you know set up. Sorry, it's 71 00:05:39.279 --> 00:05:41.800 not in the it's not in the system. I said, you know, 72 00:05:41.879 --> 00:05:45.720 as your manager here, and he's you know, we went through the rigmarole. 73 00:05:46.040 --> 00:05:50.550 Well, I reached out to their online customer service and I took photographs 74 00:05:50.589 --> 00:05:53.709 of it. I said look, you know, this is a coat hanging 75 00:05:53.790 --> 00:05:58.430 on a act that's clearly marked fifty percent off, and I got to say 76 00:05:58.470 --> 00:06:03.459 they're their customer service department failed utterly. You know, they were unable to 77 00:06:03.699 --> 00:06:06.459 help me at all and in fact, they said, you know, you 78 00:06:06.579 --> 00:06:10.980 really need to talk to the store manager at that individual store because we're just 79 00:06:11.100 --> 00:06:14.420 the online store. And I said, well, you know, I talked 80 00:06:14.459 --> 00:06:16.699 to him and he wouldn't honor the sign that was in his own store that 81 00:06:16.819 --> 00:06:20.170 clearly said fifty percent off, and and they said, well, you'll you 82 00:06:20.290 --> 00:06:24.569 really have to talk to him or his regional manager. And I said, 83 00:06:24.769 --> 00:06:27.290 well, who's the regional manager? And they said, I don't know, 84 00:06:28.810 --> 00:06:31.730 I don't know. This is so it just sound like talking to Columbia. 85 00:06:31.769 --> 00:06:35.399 I'm talking to the actual corporation. WHO's the regional manager? Who Do I 86 00:06:35.480 --> 00:06:41.000 askalate this to? I don't know. They literally failed every step of the 87 00:06:41.079 --> 00:06:46.079 way, and so what could have been an opportunity to build a positive customer 88 00:06:46.160 --> 00:06:48.550 relationship. Fifty percent off, you know, drawing a customer in. I 89 00:06:48.629 --> 00:06:51.310 was on the street, I wasn't planning on going into the store. They, 90 00:06:51.509 --> 00:06:55.029 you know, lured me into the store with their sign. They could 91 00:06:55.029 --> 00:06:58.829 have gathered information about, you know, who I was and how I was 92 00:06:58.949 --> 00:07:02.339 responding to their sales and they probably didn't even bother to figure out that I 93 00:07:02.420 --> 00:07:08.819 was angry about the fact that I had been snubbed and that they're advertising had 94 00:07:08.819 --> 00:07:13.579 been deceptive and that the store manager operating their store was, you know, 95 00:07:13.779 --> 00:07:18.850 clearly using false advertising. And so it's an opportunity lost for a company like 96 00:07:18.970 --> 00:07:24.769 Columbia. They lost my business and in fact I have three children and a 97 00:07:24.850 --> 00:07:28.290 wife who no longer shop at Columbia as a result. It was insulting. 98 00:07:28.329 --> 00:07:30.959 It was a, you know, very poor decision on their part and the 99 00:07:31.079 --> 00:07:36.240 fact that they're, you know, their organizational customer service was so bad that 100 00:07:36.319 --> 00:07:41.199 they couldn't even recognize that they'd made a mistake was humiliating for that. I 101 00:07:41.279 --> 00:07:44.750 mean, you know, it's just a blunder for any corporation to go through 102 00:07:44.790 --> 00:07:48.790 life blind to what it's doing and how it's treating consumers. So lots of 103 00:07:48.910 --> 00:07:54.509 opportunities missed and it ended up in a book right and on this show, 104 00:07:54.910 --> 00:07:57.990 on this show all over, and I'm shy, I would guess others, 105 00:07:58.310 --> 00:08:03.540 all over one hundred seventy five dollars, right, right, just so crazy. 106 00:08:03.579 --> 00:08:05.699 When you think about that long term, I just think you just, 107 00:08:05.980 --> 00:08:09.740 you know, your family alone could purchase en x that amount over the next 108 00:08:09.740 --> 00:08:13.250 five years. Oh in on ski season. Right, there you go. 109 00:08:13.610 --> 00:08:16.370 So, anyway, as soon as I saw that word, I always read 110 00:08:16.370 --> 00:08:20.410 a book with a Pencil and I underline things and put things in parentheses. 111 00:08:20.490 --> 00:08:24.009 Periodically I'll mark things in the column. So when I saw customer experience as 112 00:08:24.050 --> 00:08:26.160 like I'm going to ask him about that. So let's let's getting into the 113 00:08:26.199 --> 00:08:31.079 invisible brand very specifically. Let's start at a high level. What are you 114 00:08:31.199 --> 00:08:35.799 addressing in the book from a brand or company perspective? Like what's going on 115 00:08:35.960 --> 00:08:39.919 at a high level from a marketers or salespersons or companies are a brand's perspective? 116 00:08:39.960 --> 00:08:45.309 Yeah, so let's let's first start with that title, the invisible brand. 117 00:08:45.750 --> 00:08:50.669 I was playing off of Adam Smith's wealth of nations, where he coined 118 00:08:50.750 --> 00:08:54.820 the phrase is the invisible hand, this feature of the economy that operates to 119 00:08:56.059 --> 00:09:00.299 create wealth and in society, and I thought, you know, we're really 120 00:09:00.620 --> 00:09:05.460 onto something new, we're experiencing a brand new force in the economy and I 121 00:09:05.580 --> 00:09:09.330 had to give it a name and I thought the invisible advertise or the invisible, 122 00:09:09.889 --> 00:09:13.929 you know, marketer, and I thought what rhymes with hand, and 123 00:09:13.169 --> 00:09:16.970 then it kind of hit me between the eyes. Invisible brand. But the 124 00:09:18.330 --> 00:09:24.090 reference is very much about revealing something that is hidden and I wanted to kind 125 00:09:24.090 --> 00:09:28.240 of pull back the curtain and show people how the technology that they're using through 126 00:09:28.240 --> 00:09:33.399 their phones and their laptops and even their televisions and cars, is being used 127 00:09:33.440 --> 00:09:41.470 by marketers to gather information, personalize information delivery back to the consumer in such 128 00:09:41.509 --> 00:09:46.149 a way that they build a relationship with you so that they can get you 129 00:09:46.710 --> 00:09:52.629 to be more open to their messaging and ultimately figure out how to use artificial 130 00:09:52.710 --> 00:09:58.259 intelligence to leverage that personal interaction to persuade you and to change the way you 131 00:09:58.580 --> 00:10:03.419 act and think and what you buy. And so kind of pulling back the 132 00:10:03.580 --> 00:10:07.289 curtain on that invisible force in our lives was really the purpose of the book. 133 00:10:07.809 --> 00:10:11.129 And it's all right here. I mean you're not talking. I mean 134 00:10:11.129 --> 00:10:15.529 there are there is some you know, future look throughout, but you mean, 135 00:10:15.570 --> 00:10:18.649 this is all very, very present. So from a consumers point of 136 00:10:18.690 --> 00:10:22.279 view, or because we're all customers as well, as you already established with 137 00:10:22.399 --> 00:10:26.320 your story. They're about your Columbia experience. Talk about this a little bit 138 00:10:26.360 --> 00:10:33.759 from the customers perspective. Yeah, so from from the perspective. Most consumers 139 00:10:33.840 --> 00:10:39.029 have this kind of sensation that they're being watched and followed. They recognize that 140 00:10:39.710 --> 00:10:43.389 they've got this device sitting on their counter that listens to them, that they 141 00:10:43.509 --> 00:10:50.299 talked to. So consumers are kind of aware and they increasingly have this awareness 142 00:10:50.340 --> 00:10:56.659 of being surveiled and and even spied upon. I often say there's a fine 143 00:10:56.700 --> 00:11:01.259 line between marketing and espionage, and so where we are today is that consumers 144 00:11:01.340 --> 00:11:07.169 have this kind of sneaking suspicion that they're being manipulate. But what I was 145 00:11:07.289 --> 00:11:11.009 trying to do in the book is to explain exactly how that works and to 146 00:11:11.169 --> 00:11:16.250 equip people with kind of the knowledge that they can think both about the opportunities 147 00:11:16.330 --> 00:11:20.440 that that this technology presents as well as the risks. And one of the 148 00:11:20.799 --> 00:11:24.960 examples, I don't think I wrote it exactly this way in the book, 149 00:11:24.960 --> 00:11:30.120 but it's kind of a simple explanation. Imagine your operating a pet store and 150 00:11:30.720 --> 00:11:33.269 somebody comes to leave their pet with you. You have the right to, 151 00:11:33.629 --> 00:11:37.230 you know, ask for specific types of information. I need to know where 152 00:11:37.230 --> 00:11:39.269 you're going, I need to know how long you're going to be there, 153 00:11:39.309 --> 00:11:43.029 I need to know how to reach you if, you know, fido gets 154 00:11:43.110 --> 00:11:48.299 sick. So there's this sharing of information where the consumer gives up certain data 155 00:11:48.980 --> 00:11:54.139 that the business owner needs in order for them to have a positive relationship. 156 00:11:54.259 --> 00:12:00.419 And consumers for the most part, believe in that kind of information exchange. 157 00:12:00.460 --> 00:12:05.330 Its transparent, it's open, they recognize that there's a value that they're getting 158 00:12:05.370 --> 00:12:07.850 back, and in this case it's the safety of my pet. But where 159 00:12:09.090 --> 00:12:11.690 we are, you know, ready to draw the line is when we move 160 00:12:11.929 --> 00:12:16.480 from using the information the way we expected to be used to doing something else 161 00:12:16.559 --> 00:12:20.200 with it. You know, in the extreme example here would be you know, 162 00:12:20.399 --> 00:12:22.360 you know I'm going to be out of town this weekend, so you 163 00:12:22.480 --> 00:12:26.759 go rob my house. Okay, now we've crossed the creepy line. I 164 00:12:26.840 --> 00:12:31.629 don't want you roaming around in my house and eating my food and, you 165 00:12:31.750 --> 00:12:37.350 know, stealing my stuff because you've been privileged to receive information about my whereabouts. 166 00:12:37.750 --> 00:12:41.750 And I think that for most consumers they sense this intuitively, that there's 167 00:12:41.789 --> 00:12:48.779 this line that business shouldn't cross. But I think where we have to understand 168 00:12:48.899 --> 00:12:52.379 that, you know, the technology has taken us is that we've gone way 169 00:12:52.419 --> 00:12:58.289 faster than regulatory environments have kept up with that we are in a realm where 170 00:12:58.330 --> 00:13:03.490 data is being transmitted about who we are, where we are, what we 171 00:13:03.610 --> 00:13:07.009 do, what we buy, what we think, and that ownership over that 172 00:13:07.450 --> 00:13:15.279 knowledge is something that there's a an active conversation going on today about, you 173 00:13:15.360 --> 00:13:18.000 know, who really owns that data and knowledge, and some people, Tim 174 00:13:18.080 --> 00:13:22.279 Burners Lee, the father of the Internet, are rather the World Wide Web, 175 00:13:22.799 --> 00:13:26.879 has said openly, you know, we need to have a Magna Carta 176 00:13:26.000 --> 00:13:31.830 for privacy, for a personal information, and I think that to a degree 177 00:13:31.909 --> 00:13:37.629 he's right that there needs to be a real serious rethinking of who owns our 178 00:13:37.750 --> 00:13:41.700 data and who owns information about us. Yeah, and we're starting to see 179 00:13:41.740 --> 00:13:46.500 that roll out a little bit. Obviously, you hit on GDPR and the 180 00:13:46.980 --> 00:13:50.419 California law that's in flight. My understanding is that several other states are kind 181 00:13:50.419 --> 00:13:54.659 of in consideration there. But you know, to your point, there's so 182 00:13:54.740 --> 00:13:58.529 much that hasn't been regulated and honestly, I'm you know, I'm a little 183 00:13:58.570 --> 00:14:01.970 bit skeptical. I wouldn't call myself, I had, a glass half empty 184 00:14:03.049 --> 00:14:07.690 type of person, but I would say on average I would not trust companies 185 00:14:07.730 --> 00:14:11.960 with their own specific motives with my data in a completely unregulated environment. The 186 00:14:13.039 --> 00:14:16.480 other interesting layer here is because, and I was happy to hear the way 187 00:14:16.480 --> 00:14:20.559 you describe the way you approach the book, because I think you really delivered 188 00:14:20.639 --> 00:14:22.639 on your own hopes or expectations for it, at least the way you describe 189 00:14:22.759 --> 00:14:24.909 here, which is, you know, peel it back and put it in 190 00:14:26.029 --> 00:14:28.149 common language. In this is how this stuff works. And so the other 191 00:14:28.269 --> 00:14:33.909 thing too, is that that data can then be paired and used and resold 192 00:14:33.990 --> 00:14:37.429 in all these other things and all of a sudden it's just like long gone 193 00:14:37.549 --> 00:14:41.379 genies out of the bottle, the Pandora's boxes open and all this informations out 194 00:14:41.419 --> 00:14:48.340 there and it's being all matched up, and now people have machine have develop 195 00:14:48.379 --> 00:14:52.500 our developing profiles that know me as well or better than myself, and not 196 00:14:52.700 --> 00:14:54.809 only, I'm getting a little bit ahead of the conversation here, not only 197 00:14:56.529 --> 00:15:00.129 predict what I'm going to do, but but prescribe it to me. So 198 00:15:00.289 --> 00:15:05.889 it is something to be sensitive to. Just stay in the spirit of let's 199 00:15:05.889 --> 00:15:09.240 make this all approachable and walk it down. Just give me a quick again 200 00:15:09.360 --> 00:15:13.320 a few terms here. Just give me some quick, lightweight definitions on them. 201 00:15:13.399 --> 00:15:20.000 Artificial intelligence. I started the book off with a Joke I've heard a 202 00:15:20.039 --> 00:15:24.350 few times, which is artificial intelligence is the art of making machines act like 203 00:15:24.549 --> 00:15:30.269 they do in the movies. And of course you know the the movie making 204 00:15:30.309 --> 00:15:35.059 industry portrays. You know, robots that can easily, you know, blend 205 00:15:35.139 --> 00:15:39.539 into society without being detected. I don't think we're there yet. I think 206 00:15:39.659 --> 00:15:43.820 that, you know, we are beyond the turing test and that you know, 207 00:15:43.980 --> 00:15:48.220 a customer Chat Bot engine can pull off being a human being, but 208 00:15:48.299 --> 00:15:52.690 I don't think we've achieved, you know, what we refer to as general 209 00:15:52.929 --> 00:15:58.570 artificial intelligence. Most of the applications of Ai that are working today are narrow 210 00:15:58.649 --> 00:16:04.120 applications. They offer specific solutions to problems that we would normally think of as 211 00:16:04.279 --> 00:16:08.840 being solved by a human but they aren't general in the sense that they don't 212 00:16:08.840 --> 00:16:12.279 solve all problems. They just solve one narrow type. So when I think 213 00:16:12.320 --> 00:16:18.830 of artificial intelligence, I think of it as the process of teaching computers to 214 00:16:18.990 --> 00:16:23.269 solve problems that we would normally think of as being solved by human beings. 215 00:16:23.309 --> 00:16:30.389 And right now the applications of AI are still narrow. They're very sophisticated and 216 00:16:30.549 --> 00:16:37.019 they're super impressive, but they're still fairly narrow in scope. Super Algorithm, 217 00:16:37.340 --> 00:16:44.860 so unlike a mathematical algorithm, computer algorithm is simply a set of instructions that 218 00:16:44.980 --> 00:16:48.610 you follow, and I in the book I talked about you have an algorithm 219 00:16:48.730 --> 00:16:52.769 for waking up in the morning. Your alarm clock goes off, you turn 220 00:16:52.809 --> 00:16:56.289 the alarm off, you swing your feet out of bed and you stand up 221 00:16:56.330 --> 00:16:59.330 and turn on the light or whatever that is. So there's a sequence of 222 00:16:59.409 --> 00:17:03.120 events that you follow that you could consider an algorithm. Well, it's interesting 223 00:17:03.159 --> 00:17:07.799 about algorithms as they can be packed into other algorithms. That little get up 224 00:17:07.799 --> 00:17:11.279 out of bed algorithm can be paired up with a brush your teeth algorithm and 225 00:17:11.400 --> 00:17:15.710 an eat breakfast algorithm and to go to work algorithm, and all of those 226 00:17:15.789 --> 00:17:18.670 can be nested inside a larger algorithm called, you know, my day, 227 00:17:19.190 --> 00:17:23.910 and the my day algorithm could be nested inside an algorithm called by month and 228 00:17:25.029 --> 00:17:29.509 my year. So you can quickly see how algorithms can pair with one another. 229 00:17:29.630 --> 00:17:33.099 They can be nested inside of one another group together to solve problems. 230 00:17:33.220 --> 00:17:37.859 So think of an algorithm. Start by thinking of an algorithm as a set 231 00:17:37.900 --> 00:17:42.220 of instructions for a computer. Follow quick follow up there on Algorithm is again 232 00:17:42.259 --> 00:17:47.690 I'm offering this for the listener who doesn't understand these things, much like myself. 233 00:17:48.250 --> 00:17:52.849 Is there a final rule in a set that basically observe, learn, 234 00:17:52.650 --> 00:18:02.240 optimize so that it's self perpetuating and gets more and more effective efficient? Yes, 235 00:18:02.400 --> 00:18:04.880 when we talk about machine learning, we're really talking about the ability for 236 00:18:06.000 --> 00:18:12.390 machines, on one hand, to detect patterns in data and use either supervised 237 00:18:12.390 --> 00:18:22.029 learning or unsupervised learning to more efficiently recognize those patterns and then ultimately learn how 238 00:18:22.109 --> 00:18:26.390 to change outcomes from those patterns. So if you see, you know, 239 00:18:26.630 --> 00:18:30.619 a shopping experience and you watch the customer decision journey from the top of the 240 00:18:30.660 --> 00:18:33.460 funnel all the way through the bottom of the funnel and you can see where 241 00:18:33.460 --> 00:18:38.259 people drop out, you can gather tens of millions of data points about the 242 00:18:38.339 --> 00:18:42.730 customer experience and you've got sixty variables to, you know, play with, 243 00:18:44.049 --> 00:18:47.809 you can start to actually see how a computer would be much more efficient at 244 00:18:47.849 --> 00:18:53.289 not only detecting patterns in that data but then learning how to actually change the 245 00:18:53.329 --> 00:18:57.359 outcomes. If I tweak this, let me watch what happens as a consumer 246 00:18:57.480 --> 00:19:02.160 moves through the funnel. If I tweak this, let me see what impact 247 00:19:02.240 --> 00:19:07.799 that has and ultimately bend the consumer experience towards more positive outcomes. Whatever your 248 00:19:07.839 --> 00:19:12.190 KPI is, whether it's a conversion are, somebody making a purchase or total 249 00:19:12.309 --> 00:19:18.430 dollars, the computer can look at all of those variables that go into the 250 00:19:18.549 --> 00:19:22.789 customer experience and then, at the end start to see, okay, if 251 00:19:22.869 --> 00:19:26.220 I, you know, do x, this happens. If I do why, 252 00:19:26.700 --> 00:19:30.940 this other thing happens, and I can start to actually bend my campaign's, 253 00:19:30.099 --> 00:19:33.980 my marketing, my customer outreach and all of those touch points with the 254 00:19:34.059 --> 00:19:41.089 consumer to change the outcome. So instead of just thinking of the modern computer 255 00:19:41.289 --> 00:19:45.410 experience as just plotting points on a graph and then drawing a line of best 256 00:19:45.490 --> 00:19:48.009 fit so that you can try to figure out, you know, where the 257 00:19:48.089 --> 00:19:51.490 next point will fall, think of it as a bunch of points on a 258 00:19:51.569 --> 00:19:56.680 graph that the computer recognizes and says, okay, if I want the next 259 00:19:56.759 --> 00:20:00.119 point to fall on this line, I have to do x, Y and 260 00:20:00.200 --> 00:20:04.519 see, and that's prescriptive. The computers starting to think about how to actually 261 00:20:04.559 --> 00:20:11.069 change the outcomes rather than merely predict the outcome. Really good. That was 262 00:20:11.109 --> 00:20:15.069 I'm so glad I asked that follow up question. Another one neural network. 263 00:20:17.630 --> 00:20:22.259 Well, the reference to neural is simply organizing a computer network in the way 264 00:20:22.539 --> 00:20:27.940 that a human brain is organized, and I think that that term is often 265 00:20:29.299 --> 00:20:33.660 a little maybe misleading and it's a you know, we're not actually building human 266 00:20:33.819 --> 00:20:37.450 brains, but what we're trying to do is we're trying to suggest that the 267 00:20:37.609 --> 00:20:42.210 interaction of the data through the network is not following a single linear path, 268 00:20:42.369 --> 00:20:48.769 that there's lots of processes that are happening simultaneously and it is the culmination of 269 00:20:48.970 --> 00:20:55.839 lots of simultaneous processes that is generating answers or outcomes, which makes it more 270 00:20:56.480 --> 00:21:00.440 of a neural network. And when we start stacking neural networks, you know, 271 00:21:00.519 --> 00:21:03.880 we can start to create what we think of as deep learning. We 272 00:21:03.960 --> 00:21:08.390 start to recognize new benefits from the complexity of the systems that were leveraging and 273 00:21:08.829 --> 00:21:15.710 that gives us some really powerful outcomes. Good last one, and this is 274 00:21:15.789 --> 00:21:21.539 a term I believe you created in constructing the book. It captures the personalized, 275 00:21:21.700 --> 00:21:26.779 persuasive ability to learn, the anthropomorphic nature of our the way the machines 276 00:21:27.259 --> 00:21:32.380 interface with us, and that is psychotechnology. Feel free to to kind of 277 00:21:32.420 --> 00:21:34.650 run with this one a little bit talk about the elements of it and how 278 00:21:34.690 --> 00:21:38.609 you arrived at that term. Yeah, in the book I define for key 279 00:21:38.730 --> 00:21:44.049 trends that you just mentioned and I'll try to illuminate them just briefly. The 280 00:21:44.170 --> 00:21:48.160 first that you mentioned and I previously mentioned is the personalization of information. If 281 00:21:48.200 --> 00:21:52.039 you and I were sitting next to each other at the airport and we both 282 00:21:52.079 --> 00:21:55.359 had our laptop so open and we went to the New York Timescom we might 283 00:21:55.440 --> 00:21:56.960 see different ads. You might see ads that are tailored to you, I 284 00:21:57.039 --> 00:22:00.670 might see ads tailored to me. So we're seeing, you know, different 285 00:22:00.750 --> 00:22:04.309 information. Your facebook feed is different than my facebook feed. The news you 286 00:22:04.430 --> 00:22:08.869 consume is tailored to you, customized to your you know, wants, and 287 00:22:10.029 --> 00:22:14.430 increasingly we are seeing information reflect back to us, almost like an echo chamber, 288 00:22:14.829 --> 00:22:18.579 of what we have previously demonstrated an interest in, and that's a real 289 00:22:18.779 --> 00:22:25.740 important fundamental idea is that the machine is now equipped to address you ethan a 290 00:22:25.980 --> 00:22:32.250 you bill in different terms. That's the personalization of information and that's distinct from 291 00:22:32.289 --> 00:22:37.289 the world I grew up in, where a broadcast tower delivered a message to 292 00:22:37.410 --> 00:22:41.569 everyone at the same time. Now computers are developing one on one relationships with 293 00:22:41.730 --> 00:22:47.799 individuals. We are even able to write news copy headlines on the fly, 294 00:22:48.640 --> 00:22:52.359 tailored to you and how you are motivated and what things you like to read. 295 00:22:52.880 --> 00:22:59.109 A second is that persuasion has become a science. We are no longer 296 00:22:59.589 --> 00:23:02.750 six people in a room with a wet finger in the air trying to feel 297 00:23:02.750 --> 00:23:04.349 out which way the wind blows. You know, it's no longer kind of 298 00:23:04.430 --> 00:23:08.630 that that simple focus group testing of messages. We are now at a point 299 00:23:08.990 --> 00:23:15.339 where we can ab test messaging across thousands and even tens of thousands, and 300 00:23:15.380 --> 00:23:19.859 if you consider Google, of giant Focus Group, billions of people and and 301 00:23:21.019 --> 00:23:26.779 the ability to persuade really comes down to defining key performance indicators that can be 302 00:23:26.900 --> 00:23:32.809 tracked, watching what messaging people are exposed to and how they behave, and 303 00:23:32.930 --> 00:23:37.170 then seeing can I change those behaviors by changing the messaging? And the answer 304 00:23:37.210 --> 00:23:41.799 is yes, we can. You know, that is something as simple as 305 00:23:41.319 --> 00:23:45.000 on your facebook APP on your phone, a little number one that pops up 306 00:23:45.400 --> 00:23:51.640 that causes you to think, hope, somebody liked my picture, somebody liked 307 00:23:51.880 --> 00:23:56.150 my post. Well, guess what? We tend to gravitate toward those things 308 00:23:56.430 --> 00:24:00.710 and compulsively click to open that up and see, well, who liked my 309 00:24:02.109 --> 00:24:06.109 who like my facebook post? That's a form of persuasion. You are being 310 00:24:06.349 --> 00:24:11.059 manipulated, you're being changed by that little number one that pops up and the 311 00:24:11.180 --> 00:24:15.220 reward you get is that little drip of dopamine in your brain that makes you 312 00:24:15.259 --> 00:24:18.740 feel like, ah, somebody loves me. Well, that kind of persuasion 313 00:24:18.819 --> 00:24:23.329 is being deployed in marketing. It's being deployed in video games. You know, 314 00:24:23.410 --> 00:24:26.809 we can keep your kids glued to their video game for seventy hours without 315 00:24:26.809 --> 00:24:32.490 going to the bathroom just to, you know, earn the the last piece 316 00:24:32.569 --> 00:24:36.690 of armor in there, you know, for their night in their medieval war 317 00:24:36.849 --> 00:24:41.039 game. So you know, we were really good at persuasion. The next 318 00:24:41.200 --> 00:24:47.480 is that natural language processing is being deployed with machine learning so that we can 319 00:24:47.599 --> 00:24:53.670 now talk to machines that learn how to persuade US using personalized information. So 320 00:24:53.710 --> 00:25:00.390 I kind of blended the last two together. There's the PERSONALIZASIAN information, persuasion 321 00:25:00.430 --> 00:25:04.230 as a science, natural language processing and machine learning all taken together, other 322 00:25:04.869 --> 00:25:11.019 create technology that we're talking to and we have to be aware that the machine 323 00:25:11.019 --> 00:25:17.099 that we're talking to is listening. It's learning about us in order to persuade 324 00:25:17.140 --> 00:25:22.049 US using our own personalized information, and I thought we need a name for 325 00:25:22.130 --> 00:25:25.250 that. We gotta what do we call that? And I thought about it 326 00:25:25.329 --> 00:25:30.690 really hard. That's technology that is operating on US psychologically. What would you 327 00:25:30.769 --> 00:25:36.680 call psychological technology? You call it psychotechnology, or you could even shorten that 328 00:25:36.720 --> 00:25:42.000 if you wanted to, psychotech. And I am committed to that idea that 329 00:25:42.160 --> 00:25:48.160 this is truly something new and unique and I need to be out there explaining 330 00:25:48.279 --> 00:25:52.230 to everybody how it works, because they deserve to know and understand how it 331 00:25:52.589 --> 00:25:56.869 is impacting them. Already. It's already at work changing you and me. 332 00:25:57.829 --> 00:26:02.910 It is, and you've already cited a couple examples of it just in that 333 00:26:03.069 --> 00:26:04.779 explanation, which is really, really good. I think the way folded those 334 00:26:04.859 --> 00:26:08.259 together made a lot of sense to me and I think it'll resonate with folks 335 00:26:08.299 --> 00:26:11.579 that kind of observe these bits and pieces. But you did is such a 336 00:26:11.619 --> 00:26:15.019 nice job pulling it all together into something that we can kind of wrap our 337 00:26:15.019 --> 00:26:18.289 heads around a little bit. So let's go just for the sake of time, 338 00:26:18.329 --> 00:26:23.009 let's let's do a little bit of a go on kind of the ethical 339 00:26:23.049 --> 00:26:26.730 considerations. You already touched on it a little bit. Trust, transparency. 340 00:26:27.329 --> 00:26:30.210 Companies need to be open, but the you know, their whole M is, 341 00:26:30.730 --> 00:26:34.160 how can I use AI to get people to buy more stuff? Right? 342 00:26:34.599 --> 00:26:38.559 And on the other side you have the customers, who you've already you've 343 00:26:38.599 --> 00:26:41.960 already kind of hinted at. This is like, I'm comfortable giving you this 344 00:26:42.119 --> 00:26:48.549 information in this context because it's an appropriate exchange of, you know, value 345 00:26:48.710 --> 00:26:53.069 my valuable information in an exchange for your valuable service or access to this document 346 00:26:53.190 --> 00:26:56.869 or whatever the case may be. I'm comfortable there, but when I get 347 00:26:56.910 --> 00:27:02.059 that little further peek behind the curtain, I get very uncomfortable. I'm creeped 348 00:27:02.099 --> 00:27:04.940 out, you know, I'm turned off, I'm kind of shocked. And 349 00:27:06.019 --> 00:27:10.779 so there's this expectation management function that it, I think, is around opening 350 00:27:10.819 --> 00:27:15.299 up the black box and being more transparent about what's going on and why. 351 00:27:15.460 --> 00:27:19.009 From the company perspective, at the same time, their motivation is to sell 352 00:27:19.089 --> 00:27:22.569 more stuff using AI. And then at a certain point, I'm getting a 353 00:27:22.569 --> 00:27:26.930 little bit farther ahead, here is, you know, you discussed it at 354 00:27:27.289 --> 00:27:32.599 couple different points in the book, this idea that we won't be able to 355 00:27:32.720 --> 00:27:37.079 recognize the machine or even the data scientists don't totally understand what's going on inside 356 00:27:37.079 --> 00:27:41.799 the black box at times, and so even that is a little bit beyond 357 00:27:41.000 --> 00:27:45.630 the company's control at some level. And so just talk about that give and 358 00:27:45.710 --> 00:27:49.750 take and the creepiness factor. But the you know, we like more targeted 359 00:27:49.789 --> 00:27:55.190 ads. Maybe just talk about that given take. Yeah, so, without 360 00:27:55.230 --> 00:27:59.660 getting too far down the kind of the Sci Fi side of this, let's 361 00:27:59.660 --> 00:28:04.339 just talk about the the consumer experience. If you are in customer service or 362 00:28:04.660 --> 00:28:10.099 if you're a business owner, if you're a marketer, I strongly, strongly 363 00:28:10.220 --> 00:28:18.490 encourage you to write this down. It's about persuasion, not coercion or deception. 364 00:28:18.009 --> 00:28:25.450 So simple formula. Most consumers are willing to be persuaded. We are 365 00:28:25.569 --> 00:28:30.680 all willing to be convinced. Convinced me, tell me that your product is 366 00:28:30.799 --> 00:28:37.960 better, but don't trick me into it, don't deceive me and don't do 367 00:28:37.160 --> 00:28:42.670 anything that is coercive, and I think that keeps you on the right side 368 00:28:42.829 --> 00:28:48.390 of the creepy line. I'll define the creepy line a little bit more for 369 00:28:48.509 --> 00:28:53.190 you as you're as you're thinking about this. There was a terrific example of 370 00:28:53.750 --> 00:29:00.140 what I consider the invisible brand at work. A number of years ago target 371 00:29:00.380 --> 00:29:03.779 had a lead scientist who is asked, can you tell us when a woman 372 00:29:03.859 --> 00:29:07.940 is pregnant? Because at a when a woman is pregnant, she makes lots 373 00:29:07.980 --> 00:29:14.769 of decisions about brands and things that she will purchase that last for decades. 374 00:29:15.049 --> 00:29:18.289 She will be you know, she raises her babies with the same kind of 375 00:29:18.769 --> 00:29:22.210 dishwashing to terge at, the same kind of toothpaste, the same kind of 376 00:29:22.849 --> 00:29:26.759 laundry to tergeon. And in the process of making those decisions she makes a 377 00:29:26.839 --> 00:29:30.160 lot of new decisions. He tries a lot of new products during the point 378 00:29:30.400 --> 00:29:36.279 where she's pregnant and then soon after when she has young children. And so 379 00:29:36.599 --> 00:29:41.029 for marketers it's a very strategically important part of a woman's life is to figure 380 00:29:41.029 --> 00:29:42.630 out how do I know when a woman is pregnant? Well, it turns 381 00:29:42.670 --> 00:29:47.549 out that target has a lot of data in their system and from your basket 382 00:29:47.549 --> 00:29:49.509 of goods, of what you're purchasing, they can actually do a pretty good 383 00:29:49.549 --> 00:29:53.819 job of determining when your pregnant. Surprisingly so, and the story goes, 384 00:29:55.019 --> 00:29:57.740 and this is now a famous kind of marketing, oneonone story. They sent 385 00:29:57.740 --> 00:30:02.859 out a mailer to women who were pregnant and one of those mailers went to 386 00:30:02.940 --> 00:30:07.059 a young woman who was sixteen years old and her father received it and was 387 00:30:07.250 --> 00:30:12.450 incensed the target would have the nerve to send his sixteen year old daughter information 388 00:30:12.529 --> 00:30:17.650 about pregnancy. She's too young, of course. And you know, surprise, 389 00:30:18.089 --> 00:30:21.839 it turned out that she was actually in fact pregnant and the algorithms that 390 00:30:21.960 --> 00:30:26.039 target knew it before dad did. And that's not the end of the story. 391 00:30:26.160 --> 00:30:29.960 The end of the story is the fact that target recognized that for that 392 00:30:30.200 --> 00:30:33.920 consumer they had crossed the creepy line. That consumer didn't need to know that 393 00:30:34.240 --> 00:30:37.869 his daughter was pregnant before he did, and the fact that a, you 394 00:30:37.950 --> 00:30:44.150 know, corporation could deduce from the basket of goods that you were purchasing that 395 00:30:44.269 --> 00:30:48.150 your daughter was pregnant was something very startling to that particular individual. And the 396 00:30:48.309 --> 00:30:55.220 learning from that that that target took away is don't be so obvious. So 397 00:30:55.460 --> 00:30:59.660 learning that they they gathered from it was, okay, we're going to put 398 00:30:59.700 --> 00:31:03.650 gas grills and golf clubs into the into the mailer so that it's not all 399 00:31:03.769 --> 00:31:07.970 just baby stuff. And oh by the way, if you're happening, to 400 00:31:07.089 --> 00:31:11.849 look past the gas grills and golf clubs and there's some you know, some 401 00:31:11.970 --> 00:31:15.450 pregnancy stuff in there. That's nice, but but please don't think that we 402 00:31:15.690 --> 00:31:19.039 sent you a pregnancy mailer or a baby mailer. So there was this degree 403 00:31:19.079 --> 00:31:26.200 of obfuscation that they started crafting into their marketing so that it wasn't so obvious 404 00:31:26.319 --> 00:31:30.759 to consumers when they were crossing the creepy line. I think where that leaves 405 00:31:30.839 --> 00:31:36.630 us today is that, as consumers, we are suspicious that we are being 406 00:31:36.869 --> 00:31:41.470 stocked and preyed upon, but we are also suspicious that the corporations are hiding 407 00:31:41.670 --> 00:31:45.630 and masking what they do. It's something I call Google Noia, which is 408 00:31:45.670 --> 00:31:51.779 kind of a combination of Google and paranoid, and this creeping Google noia that 409 00:31:51.900 --> 00:31:56.460 we experience suggest to US somehow that, you know, all of these search 410 00:31:56.579 --> 00:32:00.569 engines and all of these recommendation engines and the ads and everything is really ganging 411 00:32:00.690 --> 00:32:06.170 up on us and and stocking us and persuading us in ways that we don't 412 00:32:06.170 --> 00:32:09.170 understand. And in fact a lot of that is actually happening. That is 413 00:32:09.289 --> 00:32:14.490 the case, and for consumers the trick is you better not let me know 414 00:32:14.690 --> 00:32:20.319 it and and so for for marketers, the simple rule is persuasion, not 415 00:32:20.519 --> 00:32:23.400 coercion or deception, and I think that if you stay on that side, 416 00:32:23.440 --> 00:32:29.599 if you know and recognize the persuasion is fair, consumers are willing to be 417 00:32:29.759 --> 00:32:37.309 persuaded and that you offer a transparent value exchange your data in exchange for these 418 00:32:37.470 --> 00:32:40.670 benefits, that you, as a company, will be more often than not 419 00:32:40.869 --> 00:32:45.380 on the right side of the creepy line. I like it. It's it's 420 00:32:45.579 --> 00:32:50.019 you know, golden rule obviously applies here. Or could you justify this decision 421 00:32:50.019 --> 00:32:52.500 or this behavior to someone you love and respect, like a family member or 422 00:32:52.619 --> 00:32:57.299 close friend? You know, it's easy to sit around and make decisions in 423 00:32:57.420 --> 00:33:00.970 the company's interests inside a closed setting, but to be able to justify it 424 00:33:00.130 --> 00:33:04.329 externally and explain it to people. If you're willing to do that, then 425 00:33:04.329 --> 00:33:07.089 you're probably on the safe side of it. I really love to another that 426 00:33:07.250 --> 00:33:12.289 that that target pregnancy story would like kind of broke into the mainstream, as 427 00:33:12.369 --> 00:33:17.480 did that Google Duplex. The AI assistant scheduling a haircut appointment in the interesting 428 00:33:17.559 --> 00:33:20.799 resolve of getting you did a drive by on that in the book. The 429 00:33:20.920 --> 00:33:25.000 interesting resolve there is that Google ended up based on customer feedback or your peep, 430 00:33:25.160 --> 00:33:30.349 you know, consumer feedback. This idea that the machine will present itself 431 00:33:30.390 --> 00:33:34.869 as a machine acting on behalf of a human. Again, like this transparent 432 00:33:34.910 --> 00:33:38.349 step so as not to act as if like acting as if it's a human 433 00:33:38.390 --> 00:33:40.980 but it's not a human. It's funny. I had a conversation about live 434 00:33:42.059 --> 00:33:45.819 chat and chat bots here on the show and and what the gentleman observed was 435 00:33:45.980 --> 00:33:52.299 that so often the people will ask directly into the chat exchange, are you 436 00:33:52.420 --> 00:33:55.210 a person? You know, like where are you? You know? Are 437 00:33:55.569 --> 00:33:59.450 It? Because people want to know that too. So it's interesting. Even 438 00:33:59.490 --> 00:34:04.809 in the interactions we want to have that level of transparency. What are the 439 00:34:04.970 --> 00:34:07.650 marketing jobs of the future, and I asked this with you know, you 440 00:34:07.769 --> 00:34:13.719 talked about you already mentioned new stories that right themselves, email copy that is 441 00:34:13.800 --> 00:34:19.239 going to optimize itself, adds that create run and optimize themselves. Where does 442 00:34:19.320 --> 00:34:22.559 this leave the marketer when we go more strategic and less tactical, because a 443 00:34:22.599 --> 00:34:27.190 lot of the tactics are knocked down by the machines, which is sensible, 444 00:34:27.230 --> 00:34:30.389 again, putting humans in position to do their best work in partnership where does 445 00:34:30.429 --> 00:34:36.389 that leave the human marketer in x number of years? Yeah, so the 446 00:34:36.469 --> 00:34:43.579 definition I typically apply to marketing is anticipating consumer demand and then finding products that 447 00:34:43.739 --> 00:34:47.739 can deliver on that demand profitably, and I think when you start to unpack 448 00:34:47.940 --> 00:34:52.449 that, that's a pretty sound definition of marketing. That isn't going to change. 449 00:34:52.570 --> 00:34:57.849 What's changing are the tools, as you said, the tactics. I'll 450 00:34:57.889 --> 00:35:04.610 give you one very clear example. That's happening today and I would extol encourage 451 00:35:04.769 --> 00:35:09.280 marketers to take this very seriously. You might remember twenty years ago, you 452 00:35:09.360 --> 00:35:14.760 know, the infancy, the dawn of kind of search and search engine optimization. 453 00:35:15.039 --> 00:35:19.519 Business is really faced with the challenge and that was, do I divert 454 00:35:19.679 --> 00:35:24.469 resources into search engine optimization? Is this important? Is this a fad or 455 00:35:24.550 --> 00:35:28.510 is this something that's going to affect my business for the next, you know, 456 00:35:28.590 --> 00:35:31.070 multi decades? Then, as we know, you know that the end 457 00:35:31.070 --> 00:35:37.059 of that story is that search engine optimization has been critical and those businesses that 458 00:35:37.260 --> 00:35:40.940 chose to invest succeeded and those that chose to ignore it did so at their 459 00:35:40.980 --> 00:35:46.539 own peril. I think we are at a similar inflection point with voice based 460 00:35:46.619 --> 00:35:51.849 interfaces. We are now talking to our cars we're talking to our television as 461 00:35:51.889 --> 00:35:54.329 we're talking to our cell phones, we're talking to, you know, Alexa 462 00:35:54.409 --> 00:36:00.010 and Siri, and for businesses today they're facing a similar challenge that they did 463 00:36:00.050 --> 00:36:05.719 twenty years ago. Is this something worth investing in? Do I make my 464 00:36:05.960 --> 00:36:10.440 business visible, for audible, if you will, through voice based interfaces? 465 00:36:10.679 --> 00:36:14.199 And so I would ask you, if you're a marketer and LEA, you're 466 00:36:14.239 --> 00:36:17.789 listening to this, ask this simple question. Can My products and services be 467 00:36:19.110 --> 00:36:24.630 purchased today through voice? Go over to Alexa and try to find your product 468 00:36:24.989 --> 00:36:29.989 through Alexa. See if Ciri can locate your business, see if Ciri can 469 00:36:30.429 --> 00:36:35.659 identify the products that you have available for sale and, more importantly, can 470 00:36:35.699 --> 00:36:40.099 you actually make the purchase using voice, because there are a lot of products 471 00:36:40.179 --> 00:36:45.380 that are already being sold using voice. I can sit down in my living 472 00:36:45.420 --> 00:36:47.489 room and I can say, you know, recommend a scary movie in My 473 00:36:47.650 --> 00:36:52.690 TV will pull up scary movies and I can select when. All using voice. 474 00:36:52.090 --> 00:36:57.010 That's changing things. You're driving in your car and you say in a 475 00:36:57.090 --> 00:37:01.760 directions to the nearest Thai restaurant. That's using voice to purchase things. Increasingly, 476 00:37:02.199 --> 00:37:07.079 we are spending money with our mouths. It's easier to spend money with 477 00:37:07.199 --> 00:37:10.639 our mouths that our fingers at this point, and for businesses who are really 478 00:37:10.760 --> 00:37:15.150 thinking hard about where things are going, you've got to recognize that this is 479 00:37:15.230 --> 00:37:20.989 an opportunity for you to either capture market share or get left behind. And 480 00:37:21.429 --> 00:37:27.550 why do I bring up voice? Voice is the consumer facing edge of artificial 481 00:37:27.630 --> 00:37:31.659 intelligence at some of our world's largest and most valuable companies. Think about the 482 00:37:31.780 --> 00:37:37.500 companies that you think of as the world's most valuable brands. Probably on that 483 00:37:37.659 --> 00:37:42.900 list, you're going to say, is apple, it's Microsoft, it's certainly 484 00:37:42.940 --> 00:37:47.489 Amazon. Well, when you think about those and Google, you're naming companies 485 00:37:47.570 --> 00:37:54.809 that have voice based ecosystems. The most valuable brands in the world have Siri 486 00:37:55.170 --> 00:38:01.880 and Alexa and Google assistant and Cortana. These are voice based systems that human 487 00:38:01.960 --> 00:38:07.119 beings are interacting with at increasing ease. You know what I think about? 488 00:38:07.119 --> 00:38:10.079 A story that I tell in the book? My neighbor's four year old son 489 00:38:10.320 --> 00:38:15.190 was able to get Alexa to play music for him and I was thinking, 490 00:38:15.550 --> 00:38:19.469 you know, this is really truly amazing, because a four year old hasn't 491 00:38:19.469 --> 00:38:23.150 necessarily nerd learned how to read they can't now the gate graphical user interface, 492 00:38:23.269 --> 00:38:31.059 but here he is easily using a voice based interface to navigate a complex environment, 493 00:38:31.659 --> 00:38:35.579 in this case Alexa, and I thought, you know, that's really 494 00:38:35.659 --> 00:38:39.219 impressive. Voice is something more innate, it's something deeper in us than reading. 495 00:38:39.300 --> 00:38:42.969 We learned to speak when we're a year, year and a half, 496 00:38:43.449 --> 00:38:45.889 whereas we don't learn to read until we're maybe four, five, six years 497 00:38:45.889 --> 00:38:52.090 old. And so that interface, that voice interface, is something that marketers 498 00:38:52.409 --> 00:38:55.559 today have to start grappling with. What does it mean to your business to 499 00:38:55.639 --> 00:39:01.719 have consumers shifting their attention to an interface which is all voice based? So 500 00:39:01.840 --> 00:39:05.840 these are some of the challenges that marketers face. To bring it back to 501 00:39:05.880 --> 00:39:10.190 your question again, the strategic role of marketing hasn't changed, but the tools 502 00:39:10.230 --> 00:39:16.030 and tactics that are available as changing extremely rapidly and marketers need to be on 503 00:39:16.150 --> 00:39:21.630 top of those changes and learn to adapt. Excellent. I have a several 504 00:39:21.710 --> 00:39:23.940 more questions that I will not be asking you. I again I found the 505 00:39:23.980 --> 00:39:30.260 invisible brand to be fun, easy and very interesting and informative, and again 506 00:39:30.300 --> 00:39:34.099 I recommend it highly to to anyone that's made it this far into the interview. 507 00:39:34.099 --> 00:39:37.420 You know there's so much more in there that we can't pack into this 508 00:39:37.579 --> 00:39:43.250 conversation. So will end here where I always end, which is on our 509 00:39:43.449 --> 00:39:46.889 number one core value at bombombing here on the show, which is human relationships. 510 00:39:47.650 --> 00:39:51.489 So I would love to give you the chance to incor mentioned someone who's 511 00:39:51.489 --> 00:39:54.840 had a positive impact on your life or your career. And counter to where 512 00:39:54.840 --> 00:39:59.440 we started with that retail story, you had maybe give a mention to a 513 00:39:59.559 --> 00:40:02.320 company that you really appreciate a respect for the type of experience they're delivering for 514 00:40:02.480 --> 00:40:07.630 you as a customer. Well, I would be a fool not to thank 515 00:40:07.789 --> 00:40:10.110 the folks that helped me write the book. You know, one of the 516 00:40:10.150 --> 00:40:15.110 things about writing a book is that it is not a solitary experience. When 517 00:40:15.110 --> 00:40:17.550 I got started, I thought this is just me and a typewriter and I 518 00:40:19.070 --> 00:40:22.380 thought I had to like lock myself in a room and not, you know, 519 00:40:22.500 --> 00:40:25.619 peek out. But in the acknowledge riots in the book I listed a 520 00:40:25.659 --> 00:40:30.099 whole Bunch of folks, but very specifically I got bogged down in the writing 521 00:40:30.139 --> 00:40:35.929 process. I recognize that I don't have the temperament personality to be somebody who 522 00:40:35.929 --> 00:40:38.449 can just, you know, knock out eight hours of writing every day until 523 00:40:38.449 --> 00:40:42.889 a book is finished. That I'm you know, I find moments, flashes 524 00:40:42.929 --> 00:40:45.409 of inspiration where I'll write you three, four, five pages and then it 525 00:40:45.409 --> 00:40:50.039 might be a week or two before I get involved with it again. And 526 00:40:50.840 --> 00:40:53.519 I ended up working with a researcher, felon in Darren, who helped me 527 00:40:54.039 --> 00:40:59.199 at a level that I you know, really it unlocked my ability to get 528 00:40:59.320 --> 00:41:02.719 book done because, instead of dreading digesting a, you know, a thirty 529 00:41:02.760 --> 00:41:06.269 page study, I could hand it to him and say, you know what, 530 00:41:06.510 --> 00:41:08.590 I think this is important, I think this is something that we should 531 00:41:08.590 --> 00:41:12.389 include, but I don't want to, you know, spend a whole chapter 532 00:41:12.510 --> 00:41:15.230 on it. Maybe if you could give me two paragraphs and he would do 533 00:41:15.349 --> 00:41:19.739 the drudgery of digesting that thing and feeding me back to to two paragraphs that 534 00:41:19.780 --> 00:41:23.300 I could easily staple in or, you know, kind of mortar into the 535 00:41:23.420 --> 00:41:29.099 bricks of this larger structure. And at first I thought this is cheating, 536 00:41:29.219 --> 00:41:31.099 this is too easy, but then I realized it doesn't. You know, 537 00:41:31.219 --> 00:41:36.250 you've got to figure out what your own faults are, what your own weaknesses 538 00:41:36.369 --> 00:41:38.650 are, and you've got to be willing to reach out and get help and 539 00:41:38.889 --> 00:41:43.250 so I owe a debt of gratitude to everybody who helped me with the book. 540 00:41:43.610 --> 00:41:46.159 To the second part of your question, who's doing it? My default 541 00:41:46.199 --> 00:41:52.159 answer here is Amazon, and I say that admiringly but also with a degree 542 00:41:52.199 --> 00:41:58.199 of caution. And the admiring is that Amazon has figured out how to connect 543 00:41:58.519 --> 00:42:01.230 the top of the funnel and the bottom of the funnel in a unique way 544 00:42:01.309 --> 00:42:07.070 that facebook and Google can't, and that is that they plant the seeds at 545 00:42:07.110 --> 00:42:08.590 the top of the funnel. They, you know, have that ability to 546 00:42:08.750 --> 00:42:14.630 plant the ideas and to see the market with people who bought this also bought 547 00:42:14.670 --> 00:42:17.539 this. You know you might like this, and in fact they've made tremendous 548 00:42:17.659 --> 00:42:22.980 in roads in digital marketing in terms of the dollars that they're bringing in, 549 00:42:23.539 --> 00:42:29.179 which is very impressive. But they also have something that Google and facebook don't, 550 00:42:29.420 --> 00:42:31.530 which is the cash register. Of course, facebook and Google have to 551 00:42:31.690 --> 00:42:36.969 rely on third party data, but here we have a company, Amazon, 552 00:42:37.329 --> 00:42:39.570 that is the cash registry. You can actually make the purchase right there, 553 00:42:39.809 --> 00:42:45.360 and so what Amazon is doing is their leveraging artificial intelligence to connect, to 554 00:42:45.840 --> 00:42:52.960 create Attu aribution between what you bought and what you experience through the customer experience, 555 00:42:53.559 --> 00:43:00.550 by understanding algorithmically the steps you took along your customer experience. Through their 556 00:43:00.670 --> 00:43:07.710 technology, they're getting better at changing and and molding your purchase behavior, and 557 00:43:07.869 --> 00:43:15.059 that's what's making them so successful is their application of artificial intelligence to the problem 558 00:43:15.539 --> 00:43:21.179 of attribution. And when I say that the attribution problem is by definition, 559 00:43:21.260 --> 00:43:24.500 it's figuring out to what do I attribute this purchase? You know, you 560 00:43:24.619 --> 00:43:28.530 bought a radio ad, you bought a TV Ad, you bought a newspaper 561 00:43:28.610 --> 00:43:31.889 ad and I have no idea which one of those caused you to walk in 562 00:43:32.090 --> 00:43:38.530 today and buy dog food. But Amazon can watch that entire experience through their 563 00:43:38.690 --> 00:43:45.639 portal, and I'm talking about through their movies and through their music and through 564 00:43:45.039 --> 00:43:49.599 all of the things that you're doing when you interact with the range of you 565 00:43:49.679 --> 00:43:53.000 know Amazon prime and all of the you know all the products that they're selling. 566 00:43:53.159 --> 00:43:58.869 You know they are a market place that includes a lot of things that 567 00:43:59.030 --> 00:44:04.510 you wouldn't traditionally think of as customer experiences, and they're able to weave all 568 00:44:04.630 --> 00:44:10.019 that together algorithmically to make observations about your behaviors what you will buy next. 569 00:44:10.699 --> 00:44:17.579 That companies like facebook and Google can't see. So they've really applied artificial intelligence 570 00:44:17.659 --> 00:44:22.420 to better understanding the customer experience and I would have to cite them as kind 571 00:44:22.420 --> 00:44:27.329 of my you know, kind of top company to watch in the space. 572 00:44:27.929 --> 00:44:30.409 Great Call in a great breakdown there. I mean just folding, and I 573 00:44:30.449 --> 00:44:35.369 didn't think about it this way, but folding in what music I'm listening to 574 00:44:35.530 --> 00:44:38.599 and how often, what movies and TV shows I'm watching through prime, what 575 00:44:38.719 --> 00:44:43.000 I'm buying at a whole foods market through the APP, you know, all 576 00:44:43.039 --> 00:44:46.320 the just a personal profile they can build on me is really, really interesting. 577 00:44:46.360 --> 00:44:50.960 Besides obviously going to order this, that or the other thing, you 578 00:44:51.039 --> 00:44:54.110 know, books and whatever else, through the website directly. There's so much 579 00:44:54.190 --> 00:44:58.349 of a profile they can build. I'm going to add one shout out. 580 00:44:58.389 --> 00:45:02.269 The gentleman who brought the two of us together today is Douglas Burdette, Marketing 581 00:45:02.309 --> 00:45:07.619 Artillery and the marketing book podcast. We were both guests on his show as 582 00:45:07.699 --> 00:45:10.500 authors of books that are relevant to marketers and marketing. So shout out to 583 00:45:10.539 --> 00:45:14.900 Douglas. Thanks for bringing us together. Thank you so much for your time 584 00:45:15.019 --> 00:45:19.019 here on the show today. I enjoyed it very much and continued success to 585 00:45:19.059 --> 00:45:21.969 you with folks want to follow up on this. They want to obviously order 586 00:45:22.050 --> 00:45:23.809 the book or connect with you. What are some ways that people can take 587 00:45:23.889 --> 00:45:30.809 this conversation step farther? The simplest thing to do is say Alexa, order 588 00:45:30.090 --> 00:45:36.800 the invisible brand by William Ammerman and it will arrive at their doorstep tomorrow morning. 589 00:45:37.159 --> 00:45:39.440 But if they're not willing to do that, they can simply go to 590 00:45:39.639 --> 00:45:45.119 my website, which is double you for William Ammerman, a m erm a 591 00:45:45.320 --> 00:45:50.510 andcom, and there they can find out or about me and about the book. 592 00:45:51.230 --> 00:45:52.989 Excellent. Thank you again so much for your time. Well done on 593 00:45:53.190 --> 00:45:57.750 the book and I just really appreciate what you shared here. Thank you so 594 00:45:57.869 --> 00:46:02.309 much, Ethan. Great work. If you found this conversation interesting and valuable, 595 00:46:02.550 --> 00:46:07.579 I know that you'll enjoy his book the invisible brand. And if you're 596 00:46:07.619 --> 00:46:12.260 thinking about how to put your humans in the best position to win inside your 597 00:46:12.380 --> 00:46:17.179 company, be sure to check out rehumanize your business. How personal videos accelerate 598 00:46:17.219 --> 00:46:23.489 sales and improve customer experience. You can learn more about it by visiting Bombombcom 599 00:46:23.969 --> 00:46:31.489 forward book. That's bomb Bombcom book, or you can search for humanize your 600 00:46:31.530 --> 00:46:36.480 business at Amazon. My name is Ethan Bute and thank you for listening to 601 00:46:36.599 --> 00:46:43.800 the BB growth show. I hate it when podcasts incessantly ask their listeners for 602 00:46:43.960 --> 00:46:46.869 reviews, but I get why they do it, because reviews are enormously helpful 603 00:46:46.869 --> 00:46:50.989 when you're trying to grow a podcast audience. So here's what we decided to 604 00:46:51.030 --> 00:46:53.630 do. If you leave a review for be tob growth and apple podcasts and 605 00:46:53.829 --> 00:46:59.230 email me a screenshot of the review to James at Sweet Fish Mediacom, I'll 606 00:46:59.269 --> 00:47:01.780 send you a signed copy of my new book, content based networking, how 607 00:47:01.820 --> 00:47:06.019 to instantly connect with anyone you want to know. We get a review, 608 00:47:06.059 --> 00:47:07.219 you get a free book. We both win.