Resources

Podcast | Not So Standard Deviations Episode 100 | RStudio (2020)

Featuring: Hilary Parker & Roger Peng In episode 100 of Not So Standard Deviations, the first ever episode prepared in advance, Hilary and Roger discuss creativity, its role in data science, and how it can be fostered through conversation. Also, follow up on coffee and oat milk. About Hilary: Hilary Parker is a Data Scientist on the styling recommendations team at Stitch Fix, a personal styling service that uses a combinations of human stylists and algorithmic recommendations to help people find what they love. At Stitch Fix, she focuses on what sorts of data to collect from clients in order to optimize clothing recommendations, as well as building out prototypes of algorithms or entirely new products based on new data sources. She is also a co-founder of the Not So Standard Deviations podcast, a bi-weekly data science podcast with Roger Peng that has over half a million downloads. Their topics of discussion include the R ecosystem, recent developments in the data science and statistics field, reproducibility and the "how" of how data scientists and statisticians work. Hilary recently authored the paper Opinionated Analysis Development based on discussions from the podcast. Prior to her career in the tech field, Hilary received her PhD in Biostatistics from Johns Hopkins School of Public Health. She lives at the San Francisco Zen Center with her partner, a Soto Zen Priest. In her free time, she enjoys exploring her home of 2 years, San Francisco. About Roger: Roger D. Peng is a Professor of Biostatistics at the Johns Hopkins Bloomberg School of Public Health where his research focuses on the development of statistical methods for addressing environmental health problems. He is the author of the popular book R Programming for Data Science and nine other books on data science and statistics. He is also the co-creator of the Johns Hopkins Data Science Specialization, the Simply Statistics blog where he writes about statistics for the public, the Not So Standard Deviations podcast with Hilary Parker, and The Effort Report podcast with Elizabeth Matsui. Roger is a Fellow of the American Statistical Association and is the recipient of the Mortimer Spiegelman Award from the American Public Health Association, which honors a statistician who has made outstanding contributions to public health. He can be found on Twitter and GitHub at @rdpeng

image: thumbnail.jpg

Transcript#

This transcript was generated automatically and may contain errors.

I'm really excited to announce our last, but absolutely not the least, keynote speakers are Roger Ping and Hilary Parker, the hosts of Not So Standard Deviations. If you've never heard of it before, really fantastic podcast, they've never invited me to be a guest speaker, but fortunately I'm not bitter about that at all.

But seriously, I'm very long-term fans of Roger and Hilary, Roger has been a very supporter of the Tidyverse really before it was the Tidyverse, and I think his USAR 2018 keynote does a better job of kind of articulating the philosophy and the benefits of the Tidyverse than I've ever managed.

I'm also very fortunate to call Hilary a friend, Hilary is a fantastic data scientist, I've had a number of really insightful conversations with her, one recent one which led me to really understand like one of the things that I do is design, but as well as being a data scientist at Stitch Fix, Hilary, I'm also very lucky that Hilary has acted as my stylist and she actually just picked out the clothes I'm wearing today.

So I'd like you to please join me in welcoming Hilary and Roger.

The pre-show checklist

Okay, so you're going to get a little behind-the-scenes action here, because the listeners of our show don't see this, but we have to do this every time, when you're dealing with a high-stakes situation like this, safety comes first. So we have a checklist that we go through every time, so we're going to go through the checklist.

Okay, Hilary, hardwired? Yeah. I guess we don't really need that. VPN turned off? Yeah. My work VPN causes problems with our connections. Yes. Wi-Fi turned off? Oh, actually, I haven't turned off my Wi-Fi. Okay. Reboot your router? Yeah. Almost always my internet is the issue. Yeah.

We did once do a whole podcast episode where we did not record. That's right. So once that happened, like never again. I guess that's how that made it onto the checklist. Yeah. We did a blameless postmortem.

So I think we're go for launch? Yeah. Go for launch. All right. So welcome to Not-So-Standard Deviations. This is episode 100.

Introducing the podcast

I'm Roger Pang from the Johns Hopkins Bloomberg School of Public Health. And I'm here with Hilary Parker of Stitch Fix.

So this episode is our first ever live episode. It's, I guess, ironically, not going to be about R or RStudio. But it's possible we showed up at the wrong conference. But I hope you'll enjoy it.

We also did some prep. Like, unusually. Usually we kind of show up and call each other and say, I was thinking about talking about this. And so this time we actually have some materials.

So I thought there are probably a few people in the audience who haven't heard of or heard our podcast. So we thought we'd just talk a little bit about kind of why we're here, what we're doing, why we're doing this. And so I guess this started in 2015, right after the R OpenSci unconference here in San Francisco, where Hilary and I kind of caught up. And I sent her an email. I said, hey, I think the subject of the email was, I want to broadcast your opinions to the world.

Because, you know, Hilary has a few opinions. And I thought there's nothing the internet likes more than a strongly opinionated person. So I sent her an email. I said, hey, I think we could do a data science podcast. I don't think there's exactly what I'm thinking about out there already in terms of other data science podcasts. And so, you know, do you want to do it? And so she got back to me. She said yes. And I sent her a wall of text emails saying, here's what I think we should do. And then there was no response.

A week later, I apologized a lot. And it was true that my boss quit that week. So there was like some reason. But yeah, it was mostly the wall of text. And then we just went for it. For five years since. Or almost five years. So 99 episodes. And so here we are.

So we've been talking about a lot of things. And one of the things over the past five years, and one of the things that we realized as we were preparing for this keynote is that we have like a record of everything we've said. So I wanted to make sure there was a visual impact here. But I actually printed out the transcripts of every episode and leafed through them. Because it's kind of rare that you have the ability to see what you were thinking over a period of time like this, aside from kind of like a personal journal. I have joked that the podcast is a little bit of like a personal journal.

So we wanted to look back and see what themes were coming up. And we never set out with an agenda for this, although I kind of thought you did at first. But we just wanted to chat. And then what was surprising was that we actually got some places during these chats. And so part of the reason I wanted to look back through these transcripts was to see how thoughts evolved over time.

Reflecting on the other keynotes

So for example, JJ's keynote yesterday was amazing. And maybe coincidentally or not, it touched on a lot of themes that we had been discussing in terms of open source. And for those of you who are wondering, we're not going to talk about that here. Yeah, we're not going to talk about open source at all. So even though our episodes have been leading up to that, and again, we may have had some information there that emboldened those conversations. Look out for episode 101.

And especially, I mean, JJ's was really good. And we were joking about the fact that I didn't realize what a nerd he was. Or what a policy nerd he was. I told him it warmed my heart when he brought up the case history of the Delaware Supreme Court. You know, nothing more riveting than a little bit of corporate case law, you know?

And it's just so lucky that someone who had so much interest in public policy but then became a highly productive coder for many years and then circled back and brought them together. It's unusual.

And then, yeah, the second keynote with all that data visualization really spoke to me. Because at Stitch Fix, I work a lot with outfit data. There's a lot of imagery in fashion, right? And so I was really inspired by that about different ways to visualize the data.

And actually, JJ also was talking about this. The book was called Shop Class is Soulcraft. And this idea of touching the data and making the data yourself. And looking through this transcript, actually, you had mentioned Scott Zeger talking about writing the data down by hand. And so I thought that was awesome. That actually comes from John Tukey, who talks about scratching the numbers down.

And then Jenny's talk was great, too. She has an incredible ability to just kind of see that higher level of stuff after having been in the trenches for so long. Going into the detail and understanding the nitty-gritty, but then being able to zoom out and create the paradigm for it. It's not easy, so good work.

I feel like this conference has been awesome, and we have fodder for many episodes, following up on this stuff. It's such a great vibe. This is your first RStudioConf. This is my first RStudioConf, yeah, that's right.

Looking back at five years of episodes

So one of the ideas that we had for this keynote was to kind of go through our thought process for the last five years, which is made a lot easier by looking at the transcripts, listening to the audio, hearing what we've been talking about. As we did that in preparation, we noticed a number of interesting moments, of course, and kind of recurring themes.

One of the interesting moments that we had, actually early on in the podcast, was our first and only use of profanity. Maybe I just want to ask, maybe do a quick poll. Who thinks it was me, and who thinks it was Hillary?

When people use spreadsheets, they're trying to avoid syntax bullshit, and data structure bullshit. Because we have a lot of that in any scripted language, where you have to know a fair amount of syntax, and get fairly facile with data structures. Thank you, Jenny Bryan.

It was a really popular episode. It was super early days, and Jenny was defending spreadsheets, essentially. Or talking about having empathy. Explaining the attraction of spreadsheets.

Another theme, one of the most controversial things I think we've ever discussed on the podcast, involves this topic here. And I want to point out that you got an oatmeal raisin for me. What is the problem? You're like the 10th person who's called me out on this. You're terrible. Why does anyone choose this? They're delicious.

Oatmeal cookies, to this day, get us a lot of feedback. And apparently there's a huge plate of them outside, ready for me to eat. I also love that episode we literally recorded in a Jimmy John's. That's right. That was, we met, I was in town for my brother's wedding, and we met in the middle between D.C. and Baltimore at a Jimmy John's and recorded it.

The other kind of, one recurring theme that has played out over multiple episodes is Hillary trying to get me to do stuff. And so one of those things is getting Netflix, apparently.

My only problem actually is that I don't have Netflix. Oh, come on. You can shell out. It's like $8 a month for Roger.

That's episode 3. Episode 28. Oh, the other piece of news that's really important is that I finally got Netflix. And 28 is like a year later, right? Yeah, because we do once every two months. So it took you a while.

Episode 36. So if you have access to HBO, highly recommend. Oh, HBO. You're like, I have to buy another service? I bought Netflix for you. Yeah, this one's an HBO exclusive. It's $14.99 a month. So we went from $8 to $14.99.

Episode 6, you say, I love movies. Episode 13, you say, I'm a film nerd. And then many times, you talked about this script book thing, which was predicting whether a movie would be successful based on its script. And you had algorithms for doing that. And then Walt Hickey came on, and you read his blog post about books versus movies. And you know about Steven Spielberg table reads?

I feel like I want the visual of handing you, like, here's all the times you've talked about movies. But do you not watch movies? How does this work? Are you just interested in movies intellectually and never watch them? I infer their content.

The role of design in data science

But what we really wanted to talk about was this kind of overarching theme that actually Hadley kind of alluded to, which is the idea of the role of design in data science. And we... this was something we chewed on for so long of essentially, like, how do you do data analysis?

And one thing that really struck me looking at it was that literally in episode one, we talk about it a lot. And actually, for a long time, I thought that this was why you wanted to do the podcast. It was, like, to solve this problem. I had no agenda. And it is surprising to look at that first transcript to see how many seeds were planted. Like, we actually talked about... we hit words and themes that were, like... were kind of where we got to, but I don't think we had a way of recognizing that those were going to be important at the time.

We have this habit of, like, telling people... like, giving people a bunch of choices, right? Like, you can do this, you can do that, you can do regression, you can do smoother, you know, whatever. Just pick... there's, like, five different models that you can... or strategies that you can implement. But we often don't, like, tell people how to choose between those types of things. You know, like, that's almost like... we explicitly don't do it.

You had an experience teaching a stat class where, at the end, someone essentially said to you, like, as a statistician, you... I think you did a good job of telling me things to do, and you told me things... definitely things not to do. But now how do I know what to do? Basically, like, thanks for teaching me all this stuff. I still don't know anything. You sent me down to the data set. I don't know, like, how to proceed. And this person got an A in the class. By far and away.

I think I came away with a few bullet points. You know, I had a lot more thinking to do, apparently. That was about, I guess, ten years ago now.

So I was thinking when you were saying that, is that, like, part of a successful data analysis is convincing someone of something? And that's, like... that is inherently, like, a one-on-one process. So we touched on, like, this... like, what does success look like? Yeah, like, that was kind of what this woman was getting at. It's like, okay, I have to do something. I have to accomplish something. Like, what is that thing? And how do I decide how to do it?

I think it's this human element that, you know, that is missing, I think, from a lot of, kind of, talk about data analysis. It's hard. And actually, that gets into, like, the next topic of kind of, like, who's building successful tools. And I think it has a lot to do with, like, people who genuinely have empathy with the user.

So that was empathy. But, yeah, like, these were, again, kind of things we threw out and didn't think about for a long time, but ended up being central themes to what we were talking about later.

DevOps, blameless postmortems, and data analysis

So, like, throughout, so we kind of have this, like, timeline of what this looks like, like, evolving over time. And one of the things, and, again, this was, like, four and a half years ago. I was still working at Etsy. And Etsy had this kind of amazing DevOps team. And so, for those of you who don't know, DevOps stands for developer operator. And it's this field where, essentially, it was, like, kind of taking IT to the next level.

So it was saying the people who run the website, the people who keep the servers on for the website, the people who do everything to kind of keep this, you know, shopping website up, they should also be the people developing the tools to keep the website up. And so it's, like, developer operator. And at Google, they kind of have a different word for it, site reliability engineer, but it's the same thing.

And I was really inspired by that work and kind of how they approach problems. One of the central tenets was this idea of a blameless postmortem, where if you ran into a problem, like, let's say your system failed and the website went down, there is a human tendency to, like, blame the person who wrote the code that failed or whatever. But Etsy focused a lot on, instead, doing this thing called a blameless postmortem, where you talk about, like, you frame the problem as though the system failed the operator rather than the operator failing within a good system. And then that opens up the conversation to, like, talk about iterating on the system and saying, you know, okay, this person went to work this day, said they wanted to do a good job. They did not say they wanted to take down the website. But then the system, like, allowed them to take down the website.

And so, like, based on this kind of idea of, like, how do we know what to do, there were sort of these two threads that started. One was how do we decide, like, how to build the artifact, like the dashboard or the email or the report, whatever. And then there was this kind of, like, how do you build a narrative? And so I think, like, the theme of the early part of the podcast, really for the first year, was focused on that artifact question of, like, how do you build a system that doesn't fail you?

And I was really energized by, like, connecting these two things and being like, well, the way we can talk about it as a community is building systems that help you avoid errors and articulating what those errors are, putting cost with those errors, and then making design decisions for your system around that.

I think a lot of the struggle that we had was in terms of kind of identifying the constraints and identifying the frame, you know, kind of the framing of this problem. And otherwise, it's so open. Yeah, and, like, if you don't define this problem that way, from, like, kind of more first principles, you end up in these, like, nasty language wars, or you get people who are just saying, like, you should use Knitter, and, like, and you don't talk about why, and it's like, okay, when is it appropriate to use Knitter if you want to avoid these types of errors? Like, you don't want to update your data but have your report be stale, you know? Like, that's a huge one.

I think the biggest thing I realized sort of almost right after was that in some ways the sort of blameless postmortem stuff is a design process, and it's a way of almost, like, making a system to force everyone to be empathetic. Yeah, because the point of the blameless postmortem is essentially to say, you know, here, listen to the problem. Don't blame the user. Like, try to put yourself in their shoes. Like, they essentially force you. Like, you can't, like, use you statements, or, you know, like, there's a lot of kind of, like, rules of engagement. And so like, it all kind of, like, started coming together.

The design thinking book club

And the path ends at episode 63. It's the first milestone. Which was the start. Some of you, if you were listeners, will remember the seven part book club that we had for Natural Process Design Thinking. Which some liked, some thought was too long.

And all I know is, like, I got these, like, messages that are, like, I guess you took a picture of the Kindle with your phone and then, like, texted me the picture. Is that right? That is correct. And so I'm, like, reading your photo of a Kindle.

And I think what ensued was, like, a series of, like, all-caps exchanges between you and me that were, like, oh, my God, I can't believe this is, like, exactly what we're talking about. Yeah. No. It was, like, oh, well, this articulates what we've been trying to say in, like, a few sentences versus, you know, you've written a series of long blog posts. I mean, I think those are so valuable. But not concise.

So I feel... this was the moment where, you know, we were talking about these... well, I guess it was... the book's called Designly Ways of Knowing. And we were kind of reading it, and she's sending me photos of the book, and I'm reading the book along with her, and I'm, like, send me the next page. And it was, like... so this was a book essentially about the theory of... like, academic theory around design and design work, and it was just, like... to us, it was this huge aha moment of, like, oh, it's all starting to really make sense.

And what was making sense is that design... the way that designers work and the way that they structure the work, the way they talk about it, you could essentially, like, swap in the word data analysis, and it was the same thing. In so many ways, yeah. Like, the way that designers talk about getting briefs, where it's, like, oh, you get a design brief, and then the... you don't just... as a designer, you don't just do what the person asks you, because that will be, like, a sub-part. The person doesn't know what they want. Like, they have figured out some way to articulate it, but your job as a designer is to, like, zoom out and be, like, identify what they actually need, and then make the best solution about that.

And that's the same... I mean, I think about that all the time with, like, someone asks you for a number, you know? They're, like, can you get me this number? And then you have to, like, as a data analyst, you have to zoom out and be, like, no, okay, what problem are you trying to solve? Why don't you catch me up on what you're doing, and maybe I'm going to solve this in a totally different way. Some of my collaborators, I think, would want me to just give them the number. I know, yeah. They don't stay collaborators for long.

And what was making sense is that design... the way that designers work and the way that they structure the work, the way they talk about it, you could essentially, like, swap in the word data analysis, and it was the same thing.

But, so, that's kind of, like, the end of our audio clips. But, like, essentially we got to this place where it's, like, okay, data analysis is a type of, essentially, like, either an independent thing or a type of design thinking, where it's, like, the answer to the student's question is essentially, like, adopt this whole other way of, like, approaching the world that's totally different than science. And we haven't, like, totally, as a field, we haven't totally, like, addressed that difference.

I kind of wonder what I would have told that student now, you know, more than 10 years later. And, you know, I think the way that I kind of think about it is that when you go outside, you don't see a data analysis walking around, right? It doesn't naturally occur like a tree or, you know, what else naturally occurs? A rock, right? So if it doesn't naturally occur, it has to be built by someone, or it has to be designed by someone, it has to be built. And so why shouldn't we use the same ideas there for a data analysis as for, you might, for a chair or a bridge or whatever?

And then I also think what's interesting is that whether you're building, like, a production machine learning pipeline or you're building an analysis, like, the tools are different, the types of testing you'll do is different, like, kind of, like, the technical requirements are different, but ultimately you're doing the same thing, which is, like, you're creating something that's gonna do something for someone. Like, either it's gonna be a recommender system that creates recommendations for a website or it's, like, this analysis that some sort of person's gonna consume and make a decision based on that. Might just be a PDF document? Yeah, it could be, I mean, even, like, an email, literally even an email with, like, one number in it is still, you have to decide, like, okay, this problem is only, this problem can be sufficiently addressed with an email and that person doesn't need much context and, therefore, if I just send them the email, that'll be enough to, like, convince them to do something.

And so that put us on a really interesting path of, like, okay, so if we think that data science is, like, a type of design, then how do designers work? Like, if you look at architects, how are they trained? If you look at, you know, other designers, how do they work in companies, et cetera?

And so one of the things in this design thinking book by Nigel Cross is talking a lot about, like, you just have to do it, like, over and over, and you're gonna get better every time you do it. You're gonna go through different thought processes, but you have to literally, like, exercise the part of your brain that does this constructive thinking rather than, like, the deductive thinking of science. So it's like, okay, so in data science, we kind of don't do that apprenticeship model so much. I mean, I think we wish we didn't have to. Yeah, like, we wish we could just, like, teach the stuff and be done with it, because it's a lot of work to, like, go through someone's work and tell them if it's working or not. But ultimately, like, if we kind of accept that this is a type of design and construction, then you need to be able to, like, practice it.

So we did some cool stuff with that where kind of, like, very different than the types of challenges you do with, like, Kaggle and stuff, where they're like, here's a data set. Analyze it. Instead, it was like, okay, let's say you had to build your system end to end. Like, you needed to answer some question. We did, like, commute times. Right. Yeah, my commute. So it was like, in San Francisco, it's like, I want to know exactly how many minutes and the variance in those minutes for different commute methods so I can know, like, the last possible minute that I can leave my apartment to make it to a meeting.

And so it's like, okay, how would you solve that? How would you get the data to solve that problem? How would you store it? How would you access it? How would you analyze it? How would you display it? And then how would you model it? What are the fixed effects? What are the random effects? It's amazing how just a simple kind of formulation of a problem can bring in every single aspect of design, of analysis, you know, presentation, communication.

And what I like about it, too, is that the things like how you store the data is sort of on the same par as, like, what models you choose. Like, all of those things are equal. You have to end up spending time on all of it instead of... I feel like as a field, and what I like about this conference is that, less so here, but as a field we focus so much just on the methods, and that really bothers me, in case that hasn't been clear.

And I was actually talking with someone last night about that, where I'm like, you know, in some ways like, the way that normal data science conferences are structured, it's like, everyone wants to talk about the best method, like the latest thing. And literally, in that system, you're motivated to have bad data, because it gives you more opportunity to do fancier models. And so, like, any system where it's like, oh, okay, this person could work on worse and worse data and be equally happy because they get to do all these fancy things they learned at this conference, is like, that's not solving the problem. That's like choosing not to solve the problem.

So by focusing on the whole thing, it's like, no, okay, can you hone in on exactly what data you need, rather than just making do with what might have fallen into your lap somewhere. I think one of the hardest things to do in general, but also in data science, is to kind of pull back from trying to maximize in a single dimension. And I think, to me, the way I interpreted even JJ's talk, is like, the way that corporations are structured is they maximize on a certain one dimension. And I think it's hard, that can end up with some good and some bad. And I think in data science, it's very tempting to kind of go for the optimal approach, go for the optimal method, go for the best prediction. But there are other elements as other stakeholders, there are other trade-offs to be made.

I really like that because in the second talk, she just touched it briefly at the end, but it was like, what are UX problems and what are designer problems? And then the one that was like, this is actually a design problem, was the loss function. It seems like it's just a data science problem, but that's actually the core user experience, is what loss function you use. And what system are you actually building?

I feel like it would be great if we could move to a place where instead of saying, this is the best thing, we could say, I really appreciate this set of trade-offs.

Creativity in data science

One other thing that, kind of like the end of this timeline, this has just been briefly, but essentially by getting to this place, now it opens up a whole other set of fields to look into, which is everything around creativity.

The book I'm mentioning, that we're mentioning here, is called The Creative Curve, and we had a one-episode book club on that. Learn our lesson. The idea with that was that it was someone, this guy, Alan Gannett, who essentially empirically studied creative people and how they operate. I thought it was awesome. I really liked it. It really made me feel like, oh, I can apply these principles to my data science work.

The principles were... there was consumption. Creative people will frequently... people who are engaged in movie making, they watch movies 20% of the time. You look at... just again, from empirically looking at people in these creative fields, consuming other people's work is a big part of it. You digest it, you think about it, and you iterate on it. If you want to write, you read a lot of books. If you want to write music, you listen to a lot of music.

One of the aspects that's difficult about data analysis is it's not always easy to read a lot of data analyses because they're not out there. Especially in the corporate world, data science specifically. If you have a big team, you can read whatever other people are doing, but otherwise, basically only at conferences. Or David Robinson's livestream. It's hard to consume a lot of data analysis is what I think it comes down to. I think that's a key step to becoming an expert in something.

Iteration. Iteration is this idea that you just keep going. You keep doing things. You do it over and over. Again, that is touched in the design literature too. You've got to keep doing things. Ideas are going to evolve over time. Part of why we wanted to go through those clips and put up this timeline is that even the creativity that we got from this podcast that I totally wasn't expecting. It took us four years. This timeline is over four years. The number of ways that we attacked this problem and thought about it was a lot. It never felt like it was work necessarily. It was just iterating and going on.

Then there's the community. Another one is creative communities. You look at artists. Andy Warhol had the factory where it was just a bunch of artists in a loft painting, giving feedback. With that, surrounding yourself with other creative people helps you be creative.

Coming to conferences like this, they're fun, they're energizing. I do think they embody this creative community. Especially because so many people are isolated as data scientists. Not every company has 100 data scientists. Most companies don't. Being able to get together and bounce ideas off and see what other people are doing. I guess what I'm trying to say is that is the work. It's not like this isn't productive, if that makes sense.

There was one agenda item for starting the podcast. Part of it was to produce that community of people who may be sitting in a department somewhere or in a company somewhere by themselves. They're forced to be the lone data analyst. They don't have anyone to talk to or they don't have the ability to hear what other people are working on and how they're approaching it. That's some of the most meaningful feedback I hear. When people are like, I'm alone and it's really nice to be able to hear how other people are working. That makes me feel really good.

I guess what I'm trying to say is that is the work. It's not like this isn't productive, if that makes sense.

And then feedback. Feedback. This one is definitely hard. A big part of in this creative curve book and just in general with design you actively are soliciting user feedback all the time. That's a huge part of it and you cannot be attached to what you've produced, essentially. You have to let the users tell you what to do. I think that that is extremely hard. Getting feedback is hard. Definitely not something that people are trained to do.

Maybe it's just curmudgeonly because I see it the most but I feel like, especially in academic stats and in data science in general, people do take it really personally. There's that quote that I kind of hate where it's like, photographers and statisticians both fall in love with their models. You see people really dig their heels in and it just gets personal.

What I like is that, at least with the user, I think in the design field there's a lot more focus on that accepting feedback.

Also, another thing, Hilary's personal corner, the one thing I did not expect from starting this kind of meditation practice and going down that path is that it genuinely made me better at getting empathy and getting feedback. The whole idea is dissociating yourself from who you think you are. It allows you to not feel personally threatened if someone challenges you. People take it personally when your whole identity is that you're a statistician and you're smart. If someone says your model is wrong, it's like, oh no, I'm not smart anymore. My whole identity is gone. Then you're defensive.

Figuring out ways, and for me, meditation, just figuring out ways to detach yourself from that makes you able to take feedback. Then that actually makes you better at the thing. It's kind of like a paradox. It's not like a magical thing. It's something that can be practiced. Exactly. The thing about the meditation practice that was a big paradigm shift for me was that these things aren't just fixed properties. It's not just like, oh, this person's good at getting feedback or this person has a thick skin and that's a set character trait and mine is not and that's done. It's like, no, there are ways to engage in practice that makes you more robust to that.

Again, looking at the neuroscience and Buddhism stuff, you have neuroplasticity and you can make new neural pathways. It's not just like, oh, yeah, I swear. There are ways that you can essentially change your brain. Actually, the Nigel Cross stuff goes into that. You look at people who are cab drivers and their spatial regions of their brains are much more developed.

Doing creative work is hard and it's very different than doing scientific work and it's also what we're doing. There are ways to get good at it. There's ways to practice it. There's ways to get good. Doing this whole podcast, being okay with reiterating and showing things that will be wrong in the future is like, it's okay and actually it'll make you better.

That's kind of why I feel like I've wanted to talk so much about the design thinking is because I want people to feel empowered to engage in that part of honing their craft and not feeling intimidated by it. I think probably a lot of people in this room don't think they're creative. I didn't think I was creative. It's just like identity. It's like I'm a scientist. Hillary does math and science. I want people to go through that same process I went through of opening up that side of myself and being like, this stuff all makes me better at this work and it's more fun because of it. You can do it too. Anyone can do it.

That's kind of why I feel like I've wanted to talk so much about the design thinking is because I want people to feel empowered to engage in that part of honing their craft and not feeling intimidated by it.

Coffee follow-up

Anyway, that was like sitcom ending. We do have one more. That was like big recurring theme. We have one more recurring theme that we are not open to feedback on.

Now we're bringing tea time to the internet. Exactly. I actually have tea today though. I've also gotten into coffee now. No! I know. Maybe I'll bring coffee time. That doesn't sound right.

The lead up to that was this podcast in part was I went to grad school where Roger was a professor and we had this tea time thing and we had all these conversations and enjoyed them. I'm into coffee. We can no longer have tea time.

On the x-axis is episode number. On the y-axis is number of mentions of the word coffee. It's good to have transcripts. Thank you, Stringer Package. Hillary switches to tea to coffee. Episode 36 is where I say people shouldn't talk about drinking things on podcasts. You go on a kind of long rant about listening to other podcasts and the fact that they're talking about the bourbon they drink. That doesn't belong on a podcast.

Based on feedback I was open to it. I got this mocha pot to make coffee and I literally couldn't give it away. We had a whole episode about if anyone here wanted it I'm willing to give it away and no one emailed us. No one wants the mocha pot. Last night we got some more feedback about a new device. If you don't like coffee don't talk to us. We're going to keep on going.

I did have a tie-in for this. I was thinking about it to the Soulcraft shop class where it's kind of like you just have to do stuff. I started making coffee and then I wanted to talk about it more. The more that you do something you can kind of fall in love with anything and you just have to do it a bunch and you start to care about it. Again, to the creativity it's like just start trying and you'll start to like it more and you'll start to feel more empowered to talk about it and then you'll bother everyone around you by talking about it way too much. And then you'll do a keynote at RStudioCon.

So that's our episode. Thanks for listening. And thanks to RStudioCon for having us here.