Welcome to our new podcast!
June 12, 2024

Navigating AI with Junaid Farooq

Navigating AI with Junaid Farooq

Join Sandy Estrada and Anjali Bansal in this exciting episode of "How I Met Your Data" as they welcome Junaid Farooq, a seasoned expert in data governance and strategy currently leading enterprise data initiatives at First Citizens Bank.  Junaid shar...

The player is loading ...
How I Met Your Data

Join Sandy Estrada and Anjali Bansal in this exciting episode of "How I Met Your Data" as they welcome Junaid Farooq, a seasoned expert in data governance and strategy currently leading enterprise data initiatives at First Citizens Bank.  Junaid shares his journey and insights on data democratization, data governance, and the impact of AI on data management.

Discover how First Citizens Bank navigated through significant acquisitions and the challenges they faced in integrating data and aligning on culture. Junaid also discusses the essential disciplines of data management and offers advice on staying versatile and continuously learning in the ever-evolving field of data.

Tune in for an unscripted, insightful conversation.

Chapters

00:00 - Introduction and Gratitude

01:38 - Initial Meeting and Podcast Invitation

03:17 - Introduction to Junaid’s Role at First Citizens Bank

05:22 - Pressure and Regulation in AI Implementation

07:31 - Importance of Foundational Data Practices

09:37 - Snowflake’s AI Capabilities Demo

11:47 - Integrating Data and Culture in Mergers

15:15 - Understanding Critical Data Elements (CDEs)

17:31 - Essential Skills for AI Initiatives

19:08 - Importance of Economic Concepts in Data Analysis

22:33 - Positioning the Chief Data Officer Role

25:55 - Data and Analytics Culture Embedding

26:39 - Data Democratization vs. AI Democratization

29:32 - Embracing Versatility and Continuous Learning

32:22 - Advice for Mid-Career Professionals

34:42 - Discussing Gen AI and Data Quality Solutions

Transcript
1
00:00:00,017 --> 00:00:04,037
Hi, Sandy Estrada here. Welcome back to another episode of How I Met Your Data.

2
00:00:04,257 --> 00:00:08,657
Anjali and I are thrilled to be joined this week by Junaid Farooq.

3
00:00:08,817 --> 00:00:11,617
He is a seasoned expert in data governance and strategy.

4
00:00:11,877 --> 00:00:15,237
He's currently leading enterprise data initiatives at First Citizens Bank.

5
00:00:15,457 --> 00:00:18,777
I first met Junaid in April at FEMA Boston.

6
00:00:19,057 --> 00:00:22,057
It was clear to me then that we had to have him on the podcast.

7
00:00:22,357 --> 00:00:26,037
Junaid and I had so many conversations on data democratization,

8
00:00:26,317 --> 00:00:30,797
data governance, data strategy. He left a lasting impression on me, to say the least.

9
00:00:30,957 --> 00:00:35,417
So as we go into our seventh episode, I just want to take a moment to thank you, our listeners.

10
00:00:35,697 --> 00:00:40,157
We've gotten a lot of feedback, some via email, some via LinkedIn,

11
00:00:40,517 --> 00:00:43,237
some via one-on-one meetings that I've had with folks.

12
00:00:43,397 --> 00:00:48,857
And we're just very thankful for you continuing to listen and continuing to

13
00:00:48,857 --> 00:00:51,917
just kind of be there with us in these conversations.

14
00:00:52,117 --> 00:00:55,897
And as we continue to have these conversations and record these episodes,

15
00:00:56,157 --> 00:00:59,257
I am personally growing more and more excited about our decision to keep them

16
00:00:59,257 --> 00:01:01,257
topic forward, but completely unscripted.

17
00:01:01,297 --> 00:01:05,257
And the conversation that you're going to hear today with Junaid perfectly exemplifies

18
00:01:05,257 --> 00:01:08,297
what we were looking for when we took this approach.

19
00:01:08,477 --> 00:01:12,097
We cover so many topics, and I'm not going to go through them right now.

20
00:01:12,177 --> 00:01:15,397
I'm going to let you discover those as the conversation unfolds.

21
00:01:15,397 --> 00:01:17,437
So without further delay, here's...

22
00:01:17,360 --> 00:01:38,160
Music.

23
00:01:38,001 --> 00:01:42,741
So, Janae, thank you for doing this again. Ever since I met you at FEMA in Boston.

24
00:01:43,441 --> 00:01:47,481
I was immediately, I think I came up to you after our first day hanging out

25
00:01:47,481 --> 00:01:50,301
together there and said, I would like to have you on my podcast.

26
00:01:50,521 --> 00:01:52,201
Really enjoyed just kind of

27
00:01:52,201 --> 00:01:56,241
your perspective on the topics that we were talking about at hand, right?

28
00:01:56,341 --> 00:01:59,521
Data democratization, data governance, data strategy. I mean,

29
00:01:59,541 --> 00:02:04,221
we covered a lot of different topics during the FEMA Boston event.

30
00:02:04,361 --> 00:02:07,861
So thank you again for taking the time to be here with us today.

31
00:02:08,001 --> 00:02:10,921
Yeah, thanks for having me. I similarly felt the same.

32
00:02:11,041 --> 00:02:13,981
I just loved your energy, your passion for the topic.

33
00:02:14,081 --> 00:02:18,821
And it's always fun meeting others who have the same energy and passion on the

34
00:02:18,821 --> 00:02:20,501
topic and similar views.

35
00:02:20,941 --> 00:02:25,221
Yeah, absolutely. I agree. I think your passion came through as well.

36
00:02:25,321 --> 00:02:30,161
And I appreciate that. I always find that if I'm energized, it's because of

37
00:02:30,161 --> 00:02:32,241
everybody else, not necessarily my energy.

38
00:02:32,701 --> 00:02:35,661
I'm just feeding off other people's energy. That's a plus.

39
00:02:36,121 --> 00:02:41,141
I always think we said once at our experience level, we didn't study data.

40
00:02:41,321 --> 00:02:45,041
We didn't choose this as a career. Many of us just wound up here.

41
00:02:45,181 --> 00:02:48,181
If you wound up here, it's generally because you really enjoy the topic.

42
00:02:48,361 --> 00:02:53,421
I have this joke. I say it like no one gets into the data game for the accolades and for the applause.

43
00:02:53,681 --> 00:02:58,221
It is a tough, tough business and you have to really love it to stay in it.

44
00:02:58,301 --> 00:03:02,101
If I have my data governance hat on, no one's really happy to see me, right?

45
00:03:02,361 --> 00:03:07,821
If I have a leather hat on in terms of data quality, our data is never good

46
00:03:07,821 --> 00:03:10,321
enough. No one's ever come up to me and said, you're done, Junaid.

47
00:03:10,661 --> 00:03:14,601
We've achieved our utopian state of data quality. It's a tough field and you

48
00:03:14,601 --> 00:03:17,121
have to love it to be in it. Yeah, absolutely.

49
00:03:17,581 --> 00:03:21,441
Maybe where we can start is a little introduction to you, if you could share

50
00:03:21,441 --> 00:03:23,241
with our audience a little bit about yourself.

51
00:03:23,501 --> 00:03:27,581
Sure. I'm currently at First Citizens Bank. I lead our enterprise data governance

52
00:03:27,581 --> 00:03:31,401
function, and I also lead our enterprise customer data function.

53
00:03:31,681 --> 00:03:36,581
My job's been very interesting. First Business Bank has had two major acquisitions in the last two years.

54
00:03:36,681 --> 00:03:41,461
They acquired CIT Bank in January, 2022, and they doubled in size when they.

55
00:03:42,097 --> 00:03:45,717
That. And they went from roughly $40 or $50 billion to $100 billion.

56
00:03:46,037 --> 00:03:49,837
And that put them into a category that required some regulatory scrutiny.

57
00:03:49,957 --> 00:03:53,537
So data governance and data management became a focus at that time.

58
00:03:53,657 --> 00:03:58,817
And then last year, out of nowhere, First Citizens Bank also acquired Silicon Valley Bank.

59
00:03:58,957 --> 00:04:04,897
And with that acquisition, they doubled again. So roughly $100 billion to over $200 billion.

60
00:04:05,157 --> 00:04:10,757
And quite suddenly, they're a category four bank. And so work has been exciting.

61
00:04:10,757 --> 00:04:14,037
You can imagine bringing three institutions together.

62
00:04:14,337 --> 00:04:19,717
Even in a single institution, there are so many different views and opinions on data.

63
00:04:19,957 --> 00:04:25,017
And even within the same institution, you sometimes struggle with lexicon and vocabulary.

64
00:04:25,457 --> 00:04:28,437
And then when you start assimilating two other large institutions,

65
00:04:28,597 --> 00:04:33,237
you can imagine the cultural dynamics, the varying views that everyone has on

66
00:04:33,237 --> 00:04:34,577
data and approaches to data.

67
00:04:34,877 --> 00:04:38,417
So it's been an interesting journey. It was probably very modern,

68
00:04:38,517 --> 00:04:42,137
very forward thinking, a little different than a more traditional bank.

69
00:04:42,317 --> 00:04:46,437
Yeah. One of the benefits that I think we had in acquiring Silicon Valley Bank

70
00:04:46,437 --> 00:04:50,377
is they had a very good data program in place and they were headed in the right

71
00:04:50,377 --> 00:04:51,877
direction on a lot of things.

72
00:04:52,037 --> 00:04:56,057
And so they were pointed in the right direction in terms of their data maturity journey.

73
00:04:56,237 --> 00:04:59,757
And there was a lot of synergies that we found that there were some things that

74
00:04:59,757 --> 00:05:03,737
Silicon Valley Bank was further along than we were. And then there were other

75
00:05:03,737 --> 00:05:05,577
things where we were moving further along.

76
00:05:05,817 --> 00:05:08,877
Bringing those efforts together, there's definitely a lot of synergies to be

77
00:05:08,877 --> 00:05:10,957
found. But culture is very different.

78
00:05:11,157 --> 00:05:13,597
And you know, you and I said when we were talking about data strategy,

79
00:05:13,737 --> 00:05:17,377
strategy is great because it's a type of culture, eat strategy for breakfast.

80
00:05:17,717 --> 00:05:22,377
Aligning on culture, aligning on literacy, essential to anything data related.

81
00:05:22,657 --> 00:05:27,517
So you have mentioned that you're leading data governance, politician in that

82
00:05:27,517 --> 00:05:32,497
respect, but also you're in charge of the customer data strategy as well. Well,

83
00:05:32,816 --> 00:05:39,256
How pervasive is AI? How much pressure is there within the bank in terms of

84
00:05:39,256 --> 00:05:41,516
the things you're trying to do with AI at the moment?

85
00:05:41,576 --> 00:05:45,336
Is there a ton of pressure around it or is it very measured because of the high

86
00:05:45,336 --> 00:05:48,016
regulation being part of a large financial institution?

87
00:05:48,256 --> 00:05:53,076
Yeah, I'd say the latter. First Citizens has a history of doing everything in a measured way.

88
00:05:53,236 --> 00:05:57,596
And we are generally not the first to market in these spaces.

89
00:05:57,836 --> 00:06:01,676
I think it's good to sort of observe what's happening and where things are heading.

90
00:06:01,836 --> 00:06:05,516
Do you remember a few years ago, Sandy, like people were all about blockchain,

91
00:06:05,656 --> 00:06:08,936
right? The way that we talk about Gen AI today, it was blockchain.

92
00:06:09,096 --> 00:06:14,676
And everything was going to be, there's gonna be some component of blockchain in everything.

93
00:06:15,156 --> 00:06:20,716
And like, where are we today with that, right? So definitely a very measured approach.

94
00:06:20,896 --> 00:06:25,776
I do think any institution would be foolish not to at least do some exploratory

95
00:06:25,776 --> 00:06:32,516
work in this space. My personal opinion is that this is definitely more real than blockchain.

96
00:06:32,776 --> 00:06:36,536
And I do think this is going to be a game changer. We're still very much in

97
00:06:36,536 --> 00:06:41,316
our infancy, but I do believe that this is not going to go away the way that other things have.

98
00:06:41,516 --> 00:06:47,236
Yeah. I believe that Gen AI has put a highlight on AI just in general,

99
00:06:47,376 --> 00:06:49,616
right? So even the things have been around for a while, right?

100
00:06:49,696 --> 00:06:53,216
Everybody was hot on machine learning over the last five years.

101
00:06:53,216 --> 00:06:59,596
And now it's even accelerated because a lot of organizational executives are

102
00:06:59,596 --> 00:07:03,296
coming to individuals saying, what can I do that's different and novel?

103
00:07:03,636 --> 00:07:06,476
You know, starting with Gen AI as the point of the conversation,

104
00:07:06,676 --> 00:07:10,736
but then that evolves into other topics and areas that AI can help with,

105
00:07:10,876 --> 00:07:13,356
which then brings along a lot of machine learning opportunities.

106
00:07:13,356 --> 00:07:19,176
I, for one, find that there's still this challenge of,

107
00:07:19,396 --> 00:07:23,836
yep, these are all great ideas and the technology can move quickly,

108
00:07:23,916 --> 00:07:29,696
but most places still struggle getting their house in order in terms of ensuring

109
00:07:29,696 --> 00:07:31,376
they have the right data and data quality.

110
00:07:31,943 --> 00:07:37,463
Yeah, listen, I think that there are a few fundamental things like the term

111
00:07:37,463 --> 00:07:40,583
Gen AI and the work around it sounds so exciting.

112
00:07:41,203 --> 00:07:44,283
It's hard not to get sort of get wrapped up in it.

113
00:07:44,363 --> 00:07:48,063
But this work, these initiatives, these platforms and these technologies are

114
00:07:48,063 --> 00:07:51,843
still dependent on the things that we've been working on for the last decade.

115
00:07:52,203 --> 00:07:57,023
Good data governance, good data quality, excellent metadata strategy.

116
00:07:57,163 --> 00:07:59,083
Like these are still foundational.

117
00:07:59,463 --> 00:08:05,623
And I think that these are essential. Poor data quality will skew results, will create bias.

118
00:08:05,923 --> 00:08:11,163
You know, it's interesting how we're so ready to try this and look at ways that

119
00:08:11,163 --> 00:08:16,843
we can leverage it without really understanding the criteria like needed for the foundation.

120
00:08:17,003 --> 00:08:20,743
Like, you know, what does a good model look like? How do we know when we're done training a model?

121
00:08:20,923 --> 00:08:25,423
That's a very basic question that I don't think I've heard a good answer for.

122
00:08:25,663 --> 00:08:31,383
You know, I would say we know that the content that's produced out of Gen EI isn't 100% accurate.

123
00:08:31,643 --> 00:08:35,663
And so the question is, how accurate is it? Is another question that we can't

124
00:08:35,663 --> 00:08:37,663
really quantify very easily.

125
00:08:37,843 --> 00:08:42,123
We're still in that phase of asking questions that we still don't know the answer to.

126
00:08:42,303 --> 00:08:46,203
I'd almost be astonished if people felt strong that they knew the answers to

127
00:08:46,203 --> 00:08:48,143
every question in this space.

128
00:08:48,403 --> 00:08:53,323
What I'm finding to be the case is that even if you think you're getting it

129
00:08:53,323 --> 00:08:58,483
part of the answer, the answer changes just down the line because there's a new innovation.

130
00:08:58,823 --> 00:09:04,443
There's a new way of handling it. I went to Snowflake's event last week and

131
00:09:04,443 --> 00:09:09,503
they were showing their front end to JetAI capabilities for many different models.

132
00:09:09,723 --> 00:09:14,003
And a lot of the use cases they showcased, it had citations.

133
00:09:14,003 --> 00:09:17,963
You can click on the citation and it will show you the document it found the

134
00:09:17,963 --> 00:09:21,283
answer from and the section of the document and highlight it for you.

135
00:09:21,383 --> 00:09:26,603
So So in the past, people said, well, I can't cite what my model is coming from.

136
00:09:26,683 --> 00:09:28,463
And they showed that you can.

137
00:09:28,583 --> 00:09:32,183
You can do that. There's opportunity for that kind of engagement with Gen AI,

138
00:09:32,363 --> 00:09:36,103
which is, again, moving the needle forward and answering a question that people

139
00:09:36,103 --> 00:09:37,363
couldn't answer before, right? Yeah.

140
00:09:37,687 --> 00:09:41,507
It's definitely moving in that direction. And I was super excited just to see

141
00:09:41,507 --> 00:09:44,467
that one capability that I hadn't seen anywhere else.

142
00:09:44,607 --> 00:09:48,467
That's an interesting point is like, how do you create a roadmap, right?

143
00:09:48,527 --> 00:09:52,767
Even if it's like a couple quarters, right? And say, okay, I'm going to go do something.

144
00:09:52,907 --> 00:09:55,827
And you'd be foolish to think that you'd ideate on something,

145
00:09:55,967 --> 00:09:58,887
create a pilot, define an outcome, understand the dependencies,

146
00:09:59,047 --> 00:10:01,867
and then go do it even within a year.

147
00:10:01,967 --> 00:10:06,067
Like that's a thing kind of aggressive. But if the pace of change is faster

148
00:10:06,067 --> 00:10:09,607
than that, then you're almost always playing from behind.

149
00:10:09,767 --> 00:10:11,787
You know, I'd say it's sort of tangentially related to this.

150
00:10:11,827 --> 00:10:15,747
I was talking to somebody about data, data programs, and how I've been in and

151
00:10:15,747 --> 00:10:18,767
out of so many institutions, both as an FTE and a consultant.

152
00:10:18,947 --> 00:10:23,067
I've never implemented anything the same way twice, whether it was like rolling

153
00:10:23,067 --> 00:10:28,007
out a data catalog or putting in data quality framework or even setting up governing bodies.

154
00:10:28,007 --> 00:10:32,367
And I always caution others, if they feel like they have a data governance playbook,

155
00:10:32,567 --> 00:10:34,787
there are thematic things that you will follow.

156
00:10:34,907 --> 00:10:38,327
And there are themes and principles that will guide you. But you will never

157
00:10:38,327 --> 00:10:40,547
do something the same way twice, period.

158
00:10:40,827 --> 00:10:45,127
You shouldn't, right? Every institution will have its own sort of nuanced way

159
00:10:45,127 --> 00:10:47,927
of working. And Gen AI just is an exponential.

160
00:10:48,247 --> 00:10:51,527
And now you layer in technology and the pace of change in technology.

161
00:10:52,007 --> 00:10:55,347
And I told this person that I was looking at work that I did three and a half

162
00:10:55,347 --> 00:10:58,867
years ago. I thought I can go back and leverage some of that work to what we're doing.

163
00:10:58,987 --> 00:11:01,767
In many cases, that work is obsolete today.

164
00:11:01,987 --> 00:11:06,007
The way of thinking about data, even three and a half years ago,

165
00:11:06,107 --> 00:11:09,067
is so different today and has become obsolete.

166
00:11:09,267 --> 00:11:14,087
And so that's another question that's come up, which is if we endeavor to do

167
00:11:14,087 --> 00:11:17,927
something and we create a plan to do it and it takes 12 months,

168
00:11:18,127 --> 00:11:19,847
how do we adjust for the change?

169
00:11:20,107 --> 00:11:23,067
How do we be good about seeing the next corner in the road?

170
00:11:23,287 --> 00:11:28,727
Yeah. Well, one, it makes me laugh that we're talking about playbooks and really

171
00:11:28,727 --> 00:11:32,047
not having a playbook because I really think of it more as a checklist.

172
00:11:32,347 --> 00:11:35,167
What are the handful of things we need to get right?

173
00:11:35,716 --> 00:11:40,976
And, you know, we need to adjust those criteria to meet the culture of the organization,

174
00:11:41,216 --> 00:11:42,536
where they are in their maturity.

175
00:11:42,756 --> 00:11:47,276
But I'm curious, Junaid, because you talked about multiple acquisitions over the last few years.

176
00:11:47,496 --> 00:11:53,536
So as you're onboarding your acquired bank, I guess, are you changing or really

177
00:11:53,536 --> 00:11:58,616
reflecting differences in the banks and the cultures and their expectations

178
00:11:58,616 --> 00:12:00,736
during those integration points?

179
00:12:00,736 --> 00:12:05,756
Listen, I think there's, again, like lots to be said about culture and the cultural challenges.

180
00:12:05,976 --> 00:12:10,076
You know, when we think about integrating data, our biggest challenges in simulating

181
00:12:10,076 --> 00:12:14,936
the data is that there are just different approaches that the different institutions have taken.

182
00:12:15,616 --> 00:12:20,696
Hi, Sandy here. Pardon the interruption. Junaid is about to introduce the concept

183
00:12:20,696 --> 00:12:25,896
of a CDE, and I wanted to make sure everybody was on the same page in terms of what that is.

184
00:12:26,016 --> 00:12:28,276
So a CDE is a critical data element.

185
00:12:28,576 --> 00:12:32,196
So for example, in a banking context like First Citizens Bank,

186
00:12:32,416 --> 00:12:35,316
a CDE might include something like customer account numbers,

187
00:12:35,616 --> 00:12:38,436
transactional amounts, regulatory reporting fields.

188
00:12:38,576 --> 00:12:43,936
Now, these elements are critical because errors or inconsistencies in these

189
00:12:43,936 --> 00:12:48,796
types of elements can lead to significant operation issues, compliant risks,

190
00:12:49,056 --> 00:12:50,836
or inaccurate insights.

191
00:12:51,156 --> 00:12:53,656
You can imagine what that could lead to in the context of a bank.

192
00:12:53,796 --> 00:12:57,176
So with that in mind, let's get back to our conversation with Junaid.

193
00:12:57,627 --> 00:13:01,207
We have this notion of CDEs. That's a thematic thing that most people have.

194
00:13:01,287 --> 00:13:04,587
What is the most critical data that we want to govern?

195
00:13:04,807 --> 00:13:08,567
We have an additional layer of nuance, which is let's tier our CDE.

196
00:13:08,667 --> 00:13:12,407
So you might have a CDE that is derived from other CDEs.

197
00:13:12,587 --> 00:13:16,187
And now what is the difference between a tier one and a tier two?

198
00:13:16,367 --> 00:13:22,607
Getting alignment on these things is sometimes very difficult because you don't know when to stop.

199
00:13:22,747 --> 00:13:26,767
We had one particular case where we had a CDE where if you followed it to the

200
00:13:26,767 --> 00:13:29,147
end, it had something like 53 or 63.

201
00:13:29,427 --> 00:13:35,507
And we're not going to increase our governance from one CDE to 63 times that amount, right?

202
00:13:35,547 --> 00:13:38,807
It's just not. And so the question and the debate becomes, what's the right

203
00:13:38,807 --> 00:13:41,267
number? Where do we stop? And how do we tier?

204
00:13:41,527 --> 00:13:44,127
And things like that. So it's the nuance where we find our challenges.

205
00:13:44,467 --> 00:13:47,807
Yeah. Just in talking to organizations about their critical data,

206
00:13:48,027 --> 00:13:52,327
I think one of the first challenges that we run into is really aligning on a

207
00:13:52,327 --> 00:13:54,467
definition of what critical truly means.

208
00:13:54,627 --> 00:14:00,247
Yeah. And then from there, starting to align data that fits that criteria for criticality.

209
00:14:00,367 --> 00:14:03,987
Yeah, it's like, again, it's like we know that thematically we should identify

210
00:14:03,987 --> 00:14:08,887
critical data and then the definition of criticality will vary from organization to organization.

211
00:14:09,187 --> 00:14:14,587
And what's critical to someone in reg reporting will be very different to someone

212
00:14:14,587 --> 00:14:16,047
who sits in marketing. today.

213
00:14:16,107 --> 00:14:21,167
Yeah, I love the comment you made earlier where you're not going to take a playbook

214
00:14:21,167 --> 00:14:25,827
and repeat it in the next org or even division of an organization, right?

215
00:14:25,947 --> 00:14:30,547
And apply it because it is unique for the reasons Anjali stated and then some.

216
00:14:30,587 --> 00:14:34,607
Every organization even is structured differently and you need to restructure

217
00:14:34,607 --> 00:14:37,967
how you do governance because the org structure is inherently different.

218
00:14:38,147 --> 00:14:40,807
They may have business and IT, they may have something in the middle.

219
00:14:40,867 --> 00:14:43,647
They may have a product team, an engineering team and a business team,

220
00:14:43,727 --> 00:14:45,287
right? There's no one size.

221
00:14:45,627 --> 00:14:49,247
Yeah, there's absolutely no one size fit, but you have to find that middle ground

222
00:14:49,247 --> 00:14:53,187
of having that checklist, having maybe some frameworks that you can reuse.

223
00:14:53,547 --> 00:14:57,387
You know, parts of to ensure that you're covering all the bases, right?

224
00:14:57,567 --> 00:15:00,787
At least that's the way I think about it. So, and I also agree with you that

225
00:15:00,787 --> 00:15:03,207
things are outdated as soon as you put them on paper.

226
00:15:03,387 --> 00:15:08,707
One thing that I believe is completely outdated is how we think about data quality

227
00:15:08,707 --> 00:15:11,067
within organizations and how we address it.

228
00:15:11,167 --> 00:15:15,007
So I wanted to maybe jump on that topic with you a little bit.

229
00:15:15,107 --> 00:15:19,147
Given that you're in a highly regulated environment, you're probably doing things

230
00:15:19,147 --> 00:15:23,727
a little bit of an old school way, is my guess. There's probably some legacy systems involved.

231
00:15:24,007 --> 00:15:28,747
If the organization were to move on to AI, what do you think would be critical

232
00:15:28,747 --> 00:15:30,407
to get right before you start that?

233
00:15:30,627 --> 00:15:36,807
I think that there's, without a doubt, out, there are probably three disciplines

234
00:15:36,807 --> 00:15:39,467
in data that I would say are essential.

235
00:15:39,647 --> 00:15:42,227
And again, this is debatable. You could add four or five, right?

236
00:15:42,927 --> 00:15:47,987
And I would say your metadata strategy, like your underlying metadata is crucial

237
00:15:47,987 --> 00:15:51,927
in having a really well thought out, robust.

238
00:15:52,622 --> 00:15:58,082
Data catalog. Without a doubt, good data quality. You need data that is fit for purpose.

239
00:15:58,322 --> 00:16:04,062
And you need to be really clear on what the purpose is and what the criteria of good looks like.

240
00:16:04,182 --> 00:16:06,802
So data quality is another discipline that I would call essential.

241
00:16:07,002 --> 00:16:11,062
And then your governance and oversight function, the policies and procedures

242
00:16:11,062 --> 00:16:15,142
that define how you do your metadata strategy and your data quality,

243
00:16:15,342 --> 00:16:18,442
the usage of things like AI. These are three essential.

244
00:16:18,902 --> 00:16:22,542
Without a doubt, if you're not thinking about these three, you would run into

245
00:16:22,542 --> 00:16:25,842
trouble with any initiative, especially a Gen AI initiative.

246
00:16:26,142 --> 00:16:30,722
And then when I think about other fundamentals to a Gen AI initiative,

247
00:16:31,002 --> 00:16:35,202
there are probably three things that I would tell people to be comfortable with.

248
00:16:35,282 --> 00:16:37,182
Machine learning and data science.

249
00:16:37,402 --> 00:16:41,882
You should be grounded in the basics and fundamentals in both of these.

250
00:16:42,022 --> 00:16:43,882
I would advise people to learn programming.

251
00:16:44,022 --> 00:16:47,822
Python, R, and these languages Languages are very easy to learn.

252
00:16:47,962 --> 00:16:53,502
I grew up, you know, I'm going to date myself a little bit. Like we studied C and C++ in college.

253
00:16:53,762 --> 00:16:59,002
And the amount of debugging for the older list, all the, you know, commiserating with me.

254
00:16:59,082 --> 00:17:03,422
These new languages you can honestly learn in a weekend. It's not very, very hard.

255
00:17:03,602 --> 00:17:08,522
And then the other sort of concept or skill that I tell people to develop is

256
00:17:08,522 --> 00:17:09,942
be comfortable with math.

257
00:17:10,122 --> 00:17:16,562
That's a little bit harder for some. But these algorithms rely on some foundational

258
00:17:16,562 --> 00:17:21,302
understanding of probability, statistics, some linear algebra.

259
00:17:21,582 --> 00:17:24,382
I think doing those data disciplines is foundational.

260
00:17:24,722 --> 00:17:29,082
And then really thinking about these other things and grounding yourself in

261
00:17:29,082 --> 00:17:31,462
some basic knowledge is essential.

262
00:17:31,722 --> 00:17:35,342
We talk about Gen AI without talking about all those other things.

263
00:17:35,542 --> 00:17:39,522
Grounding yourself and experience in those is essential. And depending on your

264
00:17:39,522 --> 00:17:43,242
level, a certain amount of effort or granularity you'd want to get into with those.

265
00:17:43,362 --> 00:17:46,842
But even executives like senior leaders, I would say the same thing to ground

266
00:17:46,842 --> 00:17:50,902
yourself in what it takes to write a Python script, how to leverage those libraries,

267
00:17:51,182 --> 00:17:53,902
have an understanding of probability and statistics.

268
00:17:54,122 --> 00:17:57,622
Those, I think, are the foundational sort of concepts.

269
00:17:57,942 --> 00:18:01,402
I'm sure others would have so many other things, but those are the things that

270
00:18:01,402 --> 00:18:04,122
come to my mind. Yeah, I would agree with that list.

271
00:18:04,242 --> 00:18:06,842
I also learned C++ in college.

272
00:18:07,002 --> 00:18:11,502
And after that class, I never looked at it again. But I'm jealous of the younger

273
00:18:11,502 --> 00:18:17,562
generation because the amount of hours I probably spent looking for that road semicolon.

274
00:18:18,522 --> 00:18:24,542
And right now you can go into Hugging Face and just copy or put it in there.

275
00:18:24,602 --> 00:18:27,102
And it'll be like, oh, you're missing this here. And you're not.

276
00:18:28,442 --> 00:18:33,162
Yeah, I still have PTSD whenever I have to use a semicolon in a regular session.

277
00:18:33,682 --> 00:18:38,682
It always was worse, right? You spend like literally 15, 20 minutes looking for it. Oh, yeah.

278
00:18:38,782 --> 00:18:42,622
And then your colleague will find it in two seconds, right? It's just like, can you look at this?

279
00:18:42,762 --> 00:18:45,562
And like two seconds later, they're pointing right at the line.

280
00:18:45,662 --> 00:18:49,662
The new generation definitely has it so much better.

281
00:18:50,318 --> 00:18:56,138
The amount of tools and technology and platforms that they have at their disposal

282
00:18:56,138 --> 00:18:58,058
and free, the best part of it.

283
00:18:58,138 --> 00:19:04,498
The amount of technology at our disposal today is just mind-blowing.

284
00:19:04,738 --> 00:19:07,978
I also have another element to add to your list.

285
00:19:08,178 --> 00:19:12,398
You mentioned executives being familiar with coding languages to some extent,

286
00:19:12,478 --> 00:19:16,818
just to understand the efforts that it takes for a team to facilitate the solution of a problem.

287
00:19:16,938 --> 00:19:20,258
I think that's really important. I would flip that on its head as well with

288
00:19:20,258 --> 00:19:24,558
the teams who are developing, because oftentimes those teams are making decisions

289
00:19:24,558 --> 00:19:28,958
in terms of what questions to ask of the data to find answers.

290
00:19:29,078 --> 00:19:33,138
And they're lacking those economic concepts that they need to be aware of to

291
00:19:33,138 --> 00:19:37,218
ensure they're asking the right questions in their analysis in order to find

292
00:19:37,218 --> 00:19:39,398
the right statistical model to go after.

293
00:19:39,658 --> 00:19:42,998
Just because you have an answer to a question, it doesn't mean you're asking

294
00:19:42,998 --> 00:19:45,898
the right question. That's a very, very good point.

295
00:19:46,098 --> 00:19:51,558
And you're right. I think that there is still so much learning to be done at every level.

296
00:19:51,758 --> 00:19:56,398
The people who are actually working hands-on keys probably have the biggest challenge.

297
00:19:56,718 --> 00:20:01,318
And I would say for any data person, one of the challenges that I have isn't

298
00:20:01,318 --> 00:20:04,358
those guys, isn't the team that has hands-on keys.

299
00:20:04,678 --> 00:20:07,938
I feel like they're invested and on the journey. right?

300
00:20:08,398 --> 00:20:13,478
It's the executive sponsorship, to be honest. And I feel like getting executives

301
00:20:13,478 --> 00:20:19,858
and leaders who don't have a history in data on board is to me is an age old problem.

302
00:20:20,038 --> 00:20:23,798
And I heard somebody say like the way that executives today have to have some

303
00:20:23,798 --> 00:20:27,578
foundational understanding of finance, like they have to know earnings per share.

304
00:20:27,718 --> 00:20:32,738
They have to know EBITDA. There are many of these CEOs, they're all grounded on certain mechanics.

305
00:20:33,058 --> 00:20:35,018
I heard somebody say that they're They're going to have to equally,

306
00:20:35,038 --> 00:20:40,358
to be successful CEOs, they're going to have to be equally grounded in the information

307
00:20:40,358 --> 00:20:41,858
age when it comes to data literacy.

308
00:20:42,178 --> 00:20:45,678
Yeah, I would agree with that. I would agree with that. I was in board readiness

309
00:20:45,678 --> 00:20:50,038
events and there were a number of current board of directors,

310
00:20:50,338 --> 00:20:55,838
advisory boards, individuals that they're currently on multiple wards and some public boards.

311
00:20:55,838 --> 00:20:59,698
And one of the things that they highlighted, which I could not believe,

312
00:20:59,858 --> 00:21:05,598
was the fact that most board members are actually illiterate in terms of digital

313
00:21:05,598 --> 00:21:07,738
and data and technology concepts.

314
00:21:08,395 --> 00:21:11,355
They're all over cyber because they have to be, right? That's part of risk.

315
00:21:11,535 --> 00:21:15,635
But they're unable to think about, is the organization doing the right things

316
00:21:15,635 --> 00:21:20,295
in order to ensure they're ahead of the curve, getting access to data or ensuring

317
00:21:20,295 --> 00:21:23,555
they're getting value from their data? They're just not there yet.

318
00:21:23,715 --> 00:21:27,515
And further, the CEO, they may delegate out to a CIO.

319
00:21:27,655 --> 00:21:29,635
But how do you find the right C-level

320
00:21:29,635 --> 00:21:33,895
executives for your team if you're unaware of what good looks like?

321
00:21:34,035 --> 00:21:37,755
Yeah, you know, I'm hopeful that that change is coming and it's coming quickly.

322
00:21:37,755 --> 00:21:40,555
The CDO office didn't exist, right?

323
00:21:40,695 --> 00:21:47,135
Not in the way it does even 10 years ago, 15 years ago. It was virtually non-existent 15 years ago.

324
00:21:47,295 --> 00:21:50,555
And it really started catching on probably 10 years ago.

325
00:21:50,635 --> 00:21:53,115
In the last five years is where you've seen probably a spike.

326
00:21:53,195 --> 00:21:56,835
I haven't seen any statistics on this, but I'd imagine it'd be like a hockey

327
00:21:56,835 --> 00:22:01,815
stick curve where in the last 10 years, you'd see this sort of like low number of CDOs.

328
00:22:01,915 --> 00:22:03,275
And so that's probably the first

329
00:22:03,275 --> 00:22:07,855
phase of this, which is the advent of the office and the role of CDO.

330
00:22:07,975 --> 00:22:12,735
And then I think what I've seen is that there's been often a misalignment of where the CDO sits.

331
00:22:13,055 --> 00:22:16,455
And I think that's still a debate. We probably have another podcast on.

332
00:22:16,635 --> 00:22:19,795
And so that's a whole other thing. So at least they exist.

333
00:22:20,095 --> 00:22:23,675
And then I think the other challenge has been, what does the CDO do?

334
00:22:23,875 --> 00:22:29,415
The CDO role is on average 24 months, some 22, 26 months, right?

335
00:22:29,415 --> 00:22:33,215
And so just as you're getting settled, there's a shift.

336
00:22:33,395 --> 00:22:37,235
There's a lot of that that comes into place. I have the personal view that a

337
00:22:37,235 --> 00:22:39,035
CDO should be a direct to a CEO.

338
00:22:39,495 --> 00:22:45,335
That's often not the case. If you can bring that CDO closer to the CEO and peers

339
00:22:45,335 --> 00:22:48,915
with everybody else, that's how you would accelerate the data literacy objective

340
00:22:48,915 --> 00:22:51,055
that exists at the C-suite. Right.

341
00:22:51,175 --> 00:22:56,175
I agree with you wholeheartedly on that. I like to think of it as you're either

342
00:22:56,175 --> 00:22:59,495
a capital C DO or you're not.

343
00:22:59,575 --> 00:23:02,955
And the capital C is when you're reporting to the CEO.

344
00:23:03,669 --> 00:23:08,269
And you have a seat at the table. You know, non-capital C is when you're in

345
00:23:08,269 --> 00:23:13,169
a division or you're part of a business unit or you're, because they have those as well.

346
00:23:13,289 --> 00:23:20,309
I've seen a lot of organizations where I have the CDO of this part of our region or our division.

347
00:23:20,949 --> 00:23:25,229
They really are in those positions. They don't own the technology stack.

348
00:23:25,429 --> 00:23:28,609
They don't own the engineering team. They don't sometimes don't even own the

349
00:23:28,609 --> 00:23:32,769
product side of it. And you're literally just brokering conversation and ideas

350
00:23:32,769 --> 00:23:36,889
for strategy without any of the ability to actually execute on it.

351
00:23:36,949 --> 00:23:40,569
That's a good point. And like, it begs the question, where should AI initiatives sit?

352
00:23:40,709 --> 00:23:45,849
Should they sit with a CDO or should they sit with a CIO or a CTO, right?

353
00:23:45,969 --> 00:23:50,609
Because it's actually a technology, but it's producing, Gen AI is a content

354
00:23:50,609 --> 00:23:54,029
producer, so it's producing data and it requires data.

355
00:23:54,149 --> 00:23:57,769
I've started seeing cheap AI officer roles too, right? Right.

356
00:23:57,769 --> 00:23:59,929
And maybe that's the answer. I'm not quite sure.

357
00:24:00,149 --> 00:24:02,429
It's another one of those questions where I don't know that there's a clear

358
00:24:02,429 --> 00:24:04,569
answer. Yeah, I could debate that all day.

359
00:24:04,789 --> 00:24:10,909
I think it just depends, right? It depends on how the organization defines those two roles.

360
00:24:11,229 --> 00:24:17,429
And if both roles are in charge of leveraging data to push business strategy

361
00:24:17,429 --> 00:24:20,489
in the right direction, it's the same role.

362
00:24:20,809 --> 00:24:25,669
I think that's a valid argument. That's the only way I look at it. It's the same role to me.

363
00:24:25,789 --> 00:24:29,769
You basically undress your chief data officer, the moment you hire a chief AI

364
00:24:29,769 --> 00:24:34,629
officer, because the AI officer is going to create that value within the workstream,

365
00:24:34,729 --> 00:24:37,609
within the operations, within the business.

366
00:24:37,949 --> 00:24:41,549
And a chief data officer then becomes a governance body immediately.

367
00:24:41,949 --> 00:24:45,129
To me, it just depends on what you're looking to achieve with that role.

368
00:24:45,369 --> 00:24:49,909
And maybe that's the evolution, the CDAO title, where the A is then replaced,

369
00:24:50,129 --> 00:24:54,989
where the analytics is then replaced with AI, perhaps, is where we're headed.

370
00:24:55,169 --> 00:24:58,369
I think it's going to be interesting to look back at this conversation in 24

371
00:24:58,369 --> 00:25:02,089
months and see where the answers are, right?

372
00:25:02,209 --> 00:25:04,289
And say, where is the industry headed?

373
00:25:04,529 --> 00:25:08,329
That's one of the reasons I like talking about this topic, about how just new

374
00:25:08,329 --> 00:25:12,769
it is and how we don't know the answers. I think it's one of the most exciting aspects of it.

375
00:25:12,889 --> 00:25:17,209
Yeah, well, it might be actually interesting to reflect back on this conversation.

376
00:25:17,369 --> 00:25:22,009
In about two months, I have an upcoming talk with a friend and a thought leader

377
00:25:22,009 --> 00:25:27,589
in data and analytics in July, where we talk about embedding data and analytics

378
00:25:27,589 --> 00:25:29,969
within an organization's data culture.

379
00:25:30,129 --> 00:25:34,389
And really it comes down to the responsibility of the chief data officer.

380
00:25:35,009 --> 00:25:40,149
Yet we haven't really seen our CDOs being as effective as we hope.

381
00:25:40,229 --> 00:25:43,189
And now where does AI actually sit?

382
00:25:43,349 --> 00:25:49,369
And our intention is to actually start solutioning some of the challenges that we've seen,

383
00:25:49,815 --> 00:25:55,595
with a group of CDOs as part of this discussion. I'll share back what we learn later in the summer.

384
00:25:55,775 --> 00:25:57,995
That would be interesting. You know, you got me thinking about something,

385
00:25:58,195 --> 00:26:02,595
Anjali. So, you know, Sandy and I talked about data democratization a few months ago now, I think.

386
00:26:02,675 --> 00:26:05,975
You just made me think of the concept of AI democratization.

387
00:26:06,395 --> 00:26:09,835
You know, I'm a big proponent to data democratization. I think,

388
00:26:09,835 --> 00:26:13,975
you know, there are noted challenges around governance and security and access.

389
00:26:14,315 --> 00:26:17,375
And somebody asked us the question on our panel, other than data governance,

390
00:26:17,375 --> 00:26:21,375
what's the best way to ensure appropriate use of data and data democratization?

391
00:26:21,495 --> 00:26:23,875
And I said, I know you don't want to hear that the answer is data governance,

392
00:26:23,975 --> 00:26:26,575
but the answer is data governance is what I said.

393
00:26:26,695 --> 00:26:30,975
And I think that you want to put data out there. I'm of the belief where we

394
00:26:30,975 --> 00:26:33,775
should view any institution where everybody's a data person.

395
00:26:33,935 --> 00:26:36,515
If you're an accountant and you're producing data, consuming data,

396
00:26:36,635 --> 00:26:39,315
and you're producing an earnings per share report, you're a data person.

397
00:26:39,595 --> 00:26:42,735
And so I have the belief that you should democratize data. Now the question

398
00:26:42,735 --> 00:26:44,535
is, do we democratize AI?

399
00:26:44,775 --> 00:26:49,275
And do we put AI in the hands of everybody? I'm convinced it's on the doorstep.

400
00:26:49,675 --> 00:26:52,995
At the end of the day, if you look at enterprise applications.

401
00:26:53,975 --> 00:26:57,775
They're every single enterprise application company in the world.

402
00:26:57,895 --> 00:27:02,015
There is no enterprise application company that is in the market today.

403
00:27:02,175 --> 00:27:05,155
And if unfortunately they're not doing this, they won't be in the market for long.

404
00:27:05,275 --> 00:27:08,935
But they are all infusing Gen AI capabilities into their platforms.

405
00:27:08,935 --> 00:27:14,955
I haven't seen a demo in the last six months that does not include some aspect of that.

406
00:27:15,155 --> 00:27:19,435
And if you work for a company and they're not doing this, find another job.

407
00:27:19,515 --> 00:27:22,375
I'm telling you this right now, because they're going to be left behind.

408
00:27:22,555 --> 00:27:26,755
Everyone has it. And if I think about AI governance within an enterprise,

409
00:27:27,155 --> 00:27:32,455
one of the tasks that they have is understanding how is that Gen AI used within

410
00:27:32,455 --> 00:27:36,955
an enterprise application that we now license or looking to license and ensuring

411
00:27:36,955 --> 00:27:39,695
that it's not going to lead us down a bad path, right?

412
00:27:39,855 --> 00:27:43,675
I mean, if you look at just some of the things that came up with HR functions,

413
00:27:43,915 --> 00:27:45,135
for example, that was an issue.

414
00:27:45,255 --> 00:27:49,035
Group inside, it was an issue on performance management, helping people figure

415
00:27:49,035 --> 00:27:50,435
out who to promote, those kinds of things.

416
00:27:50,515 --> 00:27:54,755
They were all issue areas that I think have evolved since those issues came

417
00:27:54,755 --> 00:27:56,915
up, I don't know, it was a year ago or so.

418
00:27:57,055 --> 00:28:00,815
But it just kind of shows every enterprise application is moving in that direction

419
00:28:00,815 --> 00:28:04,675
because they want to make the use of the application as frictionless as possible

420
00:28:04,675 --> 00:28:06,015
and user-friendly as possible.

421
00:28:06,135 --> 00:28:09,835
And there's nothing more user-friendly than a chatbot. So that's where they're headed.

422
00:28:10,035 --> 00:28:15,135
I mean, even Copilot on Microsoft, they're throwing it onto your desktop to

423
00:28:15,135 --> 00:28:17,775
the point where I don't have to go navigate to a file.

424
00:28:18,235 --> 00:28:22,475
I could just say, give me X, Y, Z, and a file opens up on my Microsoft computer,

425
00:28:22,675 --> 00:28:23,815
right? So I have an Apple.

426
00:28:23,935 --> 00:28:26,415
But anyway, you get my point? I think that's where we're headed.

427
00:28:26,655 --> 00:28:30,775
I also have seen, we're going to have another episode where we talk about the

428
00:28:30,775 --> 00:28:34,935
Snowflake conference, but I'll just add this to quick little thing I saw at

429
00:28:34,935 --> 00:28:37,695
the conference, which was they were releasing all these capabilities.

430
00:28:38,095 --> 00:28:41,375
And at one point they said, everyone should be able to build a chatbot.

431
00:28:41,535 --> 00:28:44,935
And then he kept repeating it. And then he said, we're going to grab an audience member.

432
00:28:45,035 --> 00:28:48,035
We're going to have them come up and build a chatbot. And in five minutes,

433
00:28:48,175 --> 00:28:52,655
this person who isn't a coder, and I believed her, she was in the back of the

434
00:28:52,655 --> 00:28:56,055
in the middle of the room, they called her name, she came up,

435
00:28:56,115 --> 00:28:59,375
she did what they told her, she hit the wrong button a few times,

436
00:28:59,535 --> 00:29:04,255
they messed up the demo for a second, but they built a chatbot bot off of PDFs.

437
00:29:04,255 --> 00:29:06,895
And it took her five minutes, and it worked.

438
00:29:07,095 --> 00:29:10,495
Pretty enticing demo. And I'm sitting there going, well, if anybody can do that,

439
00:29:10,615 --> 00:29:15,655
imagine the possibilities of the ability to just really ensure that all the

440
00:29:15,655 --> 00:29:19,315
knowledge within an organization is readily available to everybody within the organization.

441
00:29:19,535 --> 00:29:25,915
That's amazing, right? And it adds to the excitement of what's out there, what the potential is.

442
00:29:26,075 --> 00:29:28,055
It sounds amazing. And I think you're right.

443
00:29:28,295 --> 00:29:31,915
Institutions that don't embrace this will likely get left behind.

444
00:29:32,135 --> 00:29:38,035
Even if it's a measured approach, I think it'd be advisable to get grounded in the topic.

445
00:29:38,355 --> 00:29:42,275
Absolutely. I avoided it when it first happened as well. Because like you,

446
00:29:42,495 --> 00:29:45,095
we've all been burned by these ideas, right?

447
00:29:45,575 --> 00:29:50,815
When I graduated from college, the dot-com era busted out of my first job.

448
00:29:50,815 --> 00:29:53,835
I was laid off because those jobs were gone overnight.

449
00:29:53,995 --> 00:29:58,055
The market failed and all these startups turned into nothings, right?

450
00:29:58,155 --> 00:30:00,215
I was part of a unicorn company and

451
00:30:00,215 --> 00:30:04,235
that turned into, I think my stock was worth $2 by the time I was gone.

452
00:30:04,415 --> 00:30:08,275
So it felt very much like that.

453
00:30:08,435 --> 00:30:12,215
And I think we will see that wave of companies getting created,

454
00:30:12,335 --> 00:30:17,375
entering semi-unicorn status, and then being thrown out the door because maybe

455
00:30:17,375 --> 00:30:21,635
they were trying to build something that should be part of a different application, et cetera.

456
00:30:21,855 --> 00:30:24,255
So that's definitely going to happen, I think, in this wave.

457
00:30:24,415 --> 00:30:29,595
But what happened to me initially was, oh, here we go, another interesting flash in the pan thing.

458
00:30:29,795 --> 00:30:32,555
Even Gartner was like, this is a shiny object a year ago, right?

459
00:30:32,655 --> 00:30:37,335
And now we're all over it all I would say I have a slightly different view of

460
00:30:37,335 --> 00:30:40,975
what will happen to all these tech companies.

461
00:30:41,650 --> 00:30:47,970
My view is that I'm with you, that they sprout up and they create a bunch of very niche solutions.

462
00:30:48,390 --> 00:30:51,690
But what I've observed is that they won't necessarily go out of business the

463
00:30:51,690 --> 00:30:53,030
way that dot-coms did overnight.

464
00:30:53,210 --> 00:30:58,270
I see them just getting assimilated, like being acquired by other larger companies.

465
00:30:58,530 --> 00:31:00,590
And then I think that you'll see consolidation.

466
00:31:00,910 --> 00:31:04,790
Sometimes I think that's why they start these companies is to plan to get acquired.

467
00:31:05,090 --> 00:31:09,610
Right. And some players in the marketplace, some cloud providers out there,

468
00:31:09,690 --> 00:31:15,090
keep it nameless this for now, but they are playing smart in terms of enabling

469
00:31:15,090 --> 00:31:19,290
a platform that people can build on. I've seen it in a number of instances.

470
00:31:19,590 --> 00:31:23,430
Salesforce has always been one of those players where, as an example,

471
00:31:23,570 --> 00:31:27,150
and they've always kind of been in this marketplace of build for Salesforce,

472
00:31:27,410 --> 00:31:28,750
build an app on top of Salesforce.

473
00:31:29,310 --> 00:31:33,950
And anybody who has a startup, and I'm not saying Salesforce specifically,

474
00:31:34,270 --> 00:31:38,710
but they should look at the marketplace and say, should I make a bet on a provider

475
00:31:38,710 --> 00:31:43,590
on whether it's AWS, Microsoft, Salesforce, Snowflake, Databricks, whatever it may be.

476
00:31:43,690 --> 00:31:47,810
Should I make a bet on a provider and build my application to work specifically

477
00:31:47,810 --> 00:31:52,570
for that platform and already is integrated inherently into that platform?

478
00:31:52,810 --> 00:31:56,130
Because that provider may gobble you up. That's your exit strategy,

479
00:31:56,270 --> 00:31:59,250
right? I'm going to make it so compelling for them. And I'm going to sell to

480
00:31:59,250 --> 00:32:00,530
every single one of their clients.

481
00:32:00,690 --> 00:32:06,090
And I'm going to ensure that I make a splash every single year with them from

482
00:32:06,090 --> 00:32:08,410
a revenue standpoint so that they can't avoid me.

483
00:32:08,570 --> 00:32:12,130
And at some point, they're going to acquire you, especially if it's on their

484
00:32:12,130 --> 00:32:15,650
platform, it's fully integrated, it's kind of native to their ecosystem.

485
00:32:16,130 --> 00:32:19,210
I think that's an easy play for them to say, hey, I'm going to go ahead and

486
00:32:19,210 --> 00:32:21,270
acquire you and relieve you of this.

487
00:32:22,410 --> 00:32:28,030
I'm with you. What would be your advice to someone who's either mid-career,

488
00:32:28,130 --> 00:32:31,290
kind of later in their career, just trying to keep up with the madness?

489
00:32:31,830 --> 00:32:35,670
You know, one advice that I've given recently is, you know, I think that in

490
00:32:35,670 --> 00:32:39,590
the last 10 years, there's been sort of these nuanced roles that people fit into.

491
00:32:39,790 --> 00:32:45,010
You know, I'm of the belief that versatility is a key to success.

492
00:32:45,350 --> 00:32:51,330
Having this skill to work across data disciplines is the advice that I would

493
00:32:51,330 --> 00:32:54,290
give people coming up. I'll give like a basketball analogy.

494
00:32:55,001 --> 00:32:59,241
So in basketball, you have these positions, you have a point guard and you have a center, right?

495
00:32:59,341 --> 00:33:03,161
And historically very different, right? In terms of size and style of play.

496
00:33:03,341 --> 00:33:07,941
But the most exciting basketball and the ones that people, teams I see most

497
00:33:07,941 --> 00:33:11,901
successful is where there is this concept of positionless players,

498
00:33:12,141 --> 00:33:15,281
people who can play at any position at any given time.

499
00:33:15,281 --> 00:33:20,941
I think the future of like data professionals is someone who is so versatile

500
00:33:20,941 --> 00:33:26,021
and can be sort of positionless and have enough skill working across all these

501
00:33:26,021 --> 00:33:29,521
disciplines is the advice that I would give. And be innocent.

502
00:33:29,781 --> 00:33:32,241
That's the other piece of advice. Expect continuous learning.

503
00:33:32,381 --> 00:33:35,501
And if you aren't going to embrace it, this isn't for you.

504
00:33:35,621 --> 00:33:40,441
I wholeheartedly agree with you primarily because I see myself as someone who

505
00:33:40,441 --> 00:33:43,841
started in left field and ended up going through all these iterations.

506
00:33:43,841 --> 00:33:45,861
In my career was not a straight line.

507
00:33:46,021 --> 00:33:53,121
It was a bunch of loop-de-loops to the point where the way I look at it is you

508
00:33:53,121 --> 00:33:58,761
cannot influence change without understanding what is required of the other person's role.

509
00:33:59,081 --> 00:34:03,521
So how are you going to influence them to primarily do the right thing if you

510
00:34:03,521 --> 00:34:05,921
don't have enough understanding to push them in that direction,

511
00:34:06,081 --> 00:34:07,301
to get what you need done?

512
00:34:07,461 --> 00:34:10,961
I think that's going to be critical. And the only way to get there is to have

513
00:34:10,961 --> 00:34:15,541
the knowledge of of how those individuals have to work and what they're trying

514
00:34:15,541 --> 00:34:18,921
to achieve and what they're doing. Is that the right thing?

515
00:34:19,061 --> 00:34:23,261
Or is there something else and you should be pushing them in the right direction?

516
00:34:23,621 --> 00:34:27,201
Yeah, fair enough. I will say that I think I personally prefer a career that

517
00:34:27,201 --> 00:34:29,661
was like loop to loop versus a straight line as well.

518
00:34:29,741 --> 00:34:34,341
So I think you're fortunate. And I think it probably adds to your skill set.

519
00:34:34,761 --> 00:34:36,241
I'm very fortunate, yes.

520
00:34:36,741 --> 00:34:41,781
I didn't appreciate it until later in life. So we completely went off script on this.

521
00:34:42,521 --> 00:34:46,101
For those listeners out there, we were actually going to talk about how Gen

522
00:34:46,101 --> 00:34:47,641
AI can solve data quality issues.

523
00:34:47,901 --> 00:34:53,081
I still think it's an incredible topic. Yeah, I do think Gen AI is prime for

524
00:34:53,081 --> 00:34:54,941
solving data quality issues.

525
00:34:55,101 --> 00:34:57,641
You know, I'll say this, you know, I think I shared with you,

526
00:34:57,661 --> 00:35:00,761
I started my career as an accountant. My first job was...

527
00:35:01,078 --> 00:35:05,718
There's an intern reconciling accounting entries for inter-entry transactions.

528
00:35:06,298 --> 00:35:10,118
What it was, was an accounting problem, a data problem. And the data problems

529
00:35:10,118 --> 00:35:14,198
were missing data, format of data, just anomalies around data.

530
00:35:14,358 --> 00:35:19,258
And if you think about it, Gen AI is perfect for these kinds of problems to

531
00:35:19,258 --> 00:35:21,038
solve when you think about data cleansing.

532
00:35:21,238 --> 00:35:27,418
But yeah, maybe another podcast hyper-focused on data quality and Gen AI,

533
00:35:27,518 --> 00:35:28,838
data governance and Gen AI.

534
00:35:29,098 --> 00:35:33,138
Yeah, absolutely. I would love it. I went in last week with that in mind,

535
00:35:33,158 --> 00:35:37,878
and I actually ran around the conference floor looking at companies who are

536
00:35:37,878 --> 00:35:42,738
solving data quality issues with ML, AI, Gen AI capabilities.

537
00:35:43,558 --> 00:35:45,898
And they are many, many out there.

538
00:35:46,078 --> 00:35:50,238
It was eye-opening to see that because it's addressing an old problem in new

539
00:35:50,238 --> 00:35:53,898
and novel ways that removes some of the headaches of how to solve the problem.

540
00:35:54,038 --> 00:35:56,238
Definitely interesting conversation to be had.

541
00:35:56,578 --> 00:36:00,558
Thank you so much for having me. It was so wonderful just chatting without boundaries.

542
00:36:00,558 --> 00:36:04,418
Yeah, as advertised, I told you that when we first chatted about the episode,

543
00:36:04,558 --> 00:36:06,258
I said, hey, we're going to have a construct.

544
00:36:06,498 --> 00:36:10,938
But if we go in another direction, we'll just take it there if the conversation is good.

545
00:36:11,038 --> 00:36:15,318
And it was. So thank you again for your time, Junaid. And I hope you have a

546
00:36:15,318 --> 00:36:18,518
fabulous weekend. Thank you, Sandy. Thank you, Angela. Appreciate it.

547
00:36:18,640 --> 00:36:32,155
Music.