Transcript
1
00:00:00,017 --> 00:00:04,037
Hi, Sandy Estrada here. Welcome back to another episode of How I Met Your Data.
2
00:00:04,257 --> 00:00:08,657
Anjali and I are thrilled to be joined this week by Junaid Farooq.
3
00:00:08,817 --> 00:00:11,617
He is a seasoned expert in data governance and strategy.
4
00:00:11,877 --> 00:00:15,237
He's currently leading enterprise data initiatives at First Citizens Bank.
5
00:00:15,457 --> 00:00:18,777
I first met Junaid in April at FEMA Boston.
6
00:00:19,057 --> 00:00:22,057
It was clear to me then that we had to have him on the podcast.
7
00:00:22,357 --> 00:00:26,037
Junaid and I had so many conversations on data democratization,
8
00:00:26,317 --> 00:00:30,797
data governance, data strategy. He left a lasting impression on me, to say the least.
9
00:00:30,957 --> 00:00:35,417
So as we go into our seventh episode, I just want to take a moment to thank you, our listeners.
10
00:00:35,697 --> 00:00:40,157
We've gotten a lot of feedback, some via email, some via LinkedIn,
11
00:00:40,517 --> 00:00:43,237
some via one-on-one meetings that I've had with folks.
12
00:00:43,397 --> 00:00:48,857
And we're just very thankful for you continuing to listen and continuing to
13
00:00:48,857 --> 00:00:51,917
just kind of be there with us in these conversations.
14
00:00:52,117 --> 00:00:55,897
And as we continue to have these conversations and record these episodes,
15
00:00:56,157 --> 00:00:59,257
I am personally growing more and more excited about our decision to keep them
16
00:00:59,257 --> 00:01:01,257
topic forward, but completely unscripted.
17
00:01:01,297 --> 00:01:05,257
And the conversation that you're going to hear today with Junaid perfectly exemplifies
18
00:01:05,257 --> 00:01:08,297
what we were looking for when we took this approach.
19
00:01:08,477 --> 00:01:12,097
We cover so many topics, and I'm not going to go through them right now.
20
00:01:12,177 --> 00:01:15,397
I'm going to let you discover those as the conversation unfolds.
21
00:01:15,397 --> 00:01:17,437
So without further delay, here's...
22
00:01:17,360 --> 00:01:38,160
Music.
23
00:01:38,001 --> 00:01:42,741
So, Janae, thank you for doing this again. Ever since I met you at FEMA in Boston.
24
00:01:43,441 --> 00:01:47,481
I was immediately, I think I came up to you after our first day hanging out
25
00:01:47,481 --> 00:01:50,301
together there and said, I would like to have you on my podcast.
26
00:01:50,521 --> 00:01:52,201
Really enjoyed just kind of
27
00:01:52,201 --> 00:01:56,241
your perspective on the topics that we were talking about at hand, right?
28
00:01:56,341 --> 00:01:59,521
Data democratization, data governance, data strategy. I mean,
29
00:01:59,541 --> 00:02:04,221
we covered a lot of different topics during the FEMA Boston event.
30
00:02:04,361 --> 00:02:07,861
So thank you again for taking the time to be here with us today.
31
00:02:08,001 --> 00:02:10,921
Yeah, thanks for having me. I similarly felt the same.
32
00:02:11,041 --> 00:02:13,981
I just loved your energy, your passion for the topic.
33
00:02:14,081 --> 00:02:18,821
And it's always fun meeting others who have the same energy and passion on the
34
00:02:18,821 --> 00:02:20,501
topic and similar views.
35
00:02:20,941 --> 00:02:25,221
Yeah, absolutely. I agree. I think your passion came through as well.
36
00:02:25,321 --> 00:02:30,161
And I appreciate that. I always find that if I'm energized, it's because of
37
00:02:30,161 --> 00:02:32,241
everybody else, not necessarily my energy.
38
00:02:32,701 --> 00:02:35,661
I'm just feeding off other people's energy. That's a plus.
39
00:02:36,121 --> 00:02:41,141
I always think we said once at our experience level, we didn't study data.
40
00:02:41,321 --> 00:02:45,041
We didn't choose this as a career. Many of us just wound up here.
41
00:02:45,181 --> 00:02:48,181
If you wound up here, it's generally because you really enjoy the topic.
42
00:02:48,361 --> 00:02:53,421
I have this joke. I say it like no one gets into the data game for the accolades and for the applause.
43
00:02:53,681 --> 00:02:58,221
It is a tough, tough business and you have to really love it to stay in it.
44
00:02:58,301 --> 00:03:02,101
If I have my data governance hat on, no one's really happy to see me, right?
45
00:03:02,361 --> 00:03:07,821
If I have a leather hat on in terms of data quality, our data is never good
46
00:03:07,821 --> 00:03:10,321
enough. No one's ever come up to me and said, you're done, Junaid.
47
00:03:10,661 --> 00:03:14,601
We've achieved our utopian state of data quality. It's a tough field and you
48
00:03:14,601 --> 00:03:17,121
have to love it to be in it. Yeah, absolutely.
49
00:03:17,581 --> 00:03:21,441
Maybe where we can start is a little introduction to you, if you could share
50
00:03:21,441 --> 00:03:23,241
with our audience a little bit about yourself.
51
00:03:23,501 --> 00:03:27,581
Sure. I'm currently at First Citizens Bank. I lead our enterprise data governance
52
00:03:27,581 --> 00:03:31,401
function, and I also lead our enterprise customer data function.
53
00:03:31,681 --> 00:03:36,581
My job's been very interesting. First Business Bank has had two major acquisitions in the last two years.
54
00:03:36,681 --> 00:03:41,461
They acquired CIT Bank in January, 2022, and they doubled in size when they.
55
00:03:42,097 --> 00:03:45,717
That. And they went from roughly $40 or $50 billion to $100 billion.
56
00:03:46,037 --> 00:03:49,837
And that put them into a category that required some regulatory scrutiny.
57
00:03:49,957 --> 00:03:53,537
So data governance and data management became a focus at that time.
58
00:03:53,657 --> 00:03:58,817
And then last year, out of nowhere, First Citizens Bank also acquired Silicon Valley Bank.
59
00:03:58,957 --> 00:04:04,897
And with that acquisition, they doubled again. So roughly $100 billion to over $200 billion.
60
00:04:05,157 --> 00:04:10,757
And quite suddenly, they're a category four bank. And so work has been exciting.
61
00:04:10,757 --> 00:04:14,037
You can imagine bringing three institutions together.
62
00:04:14,337 --> 00:04:19,717
Even in a single institution, there are so many different views and opinions on data.
63
00:04:19,957 --> 00:04:25,017
And even within the same institution, you sometimes struggle with lexicon and vocabulary.
64
00:04:25,457 --> 00:04:28,437
And then when you start assimilating two other large institutions,
65
00:04:28,597 --> 00:04:33,237
you can imagine the cultural dynamics, the varying views that everyone has on
66
00:04:33,237 --> 00:04:34,577
data and approaches to data.
67
00:04:34,877 --> 00:04:38,417
So it's been an interesting journey. It was probably very modern,
68
00:04:38,517 --> 00:04:42,137
very forward thinking, a little different than a more traditional bank.
69
00:04:42,317 --> 00:04:46,437
Yeah. One of the benefits that I think we had in acquiring Silicon Valley Bank
70
00:04:46,437 --> 00:04:50,377
is they had a very good data program in place and they were headed in the right
71
00:04:50,377 --> 00:04:51,877
direction on a lot of things.
72
00:04:52,037 --> 00:04:56,057
And so they were pointed in the right direction in terms of their data maturity journey.
73
00:04:56,237 --> 00:04:59,757
And there was a lot of synergies that we found that there were some things that
74
00:04:59,757 --> 00:05:03,737
Silicon Valley Bank was further along than we were. And then there were other
75
00:05:03,737 --> 00:05:05,577
things where we were moving further along.
76
00:05:05,817 --> 00:05:08,877
Bringing those efforts together, there's definitely a lot of synergies to be
77
00:05:08,877 --> 00:05:10,957
found. But culture is very different.
78
00:05:11,157 --> 00:05:13,597
And you know, you and I said when we were talking about data strategy,
79
00:05:13,737 --> 00:05:17,377
strategy is great because it's a type of culture, eat strategy for breakfast.
80
00:05:17,717 --> 00:05:22,377
Aligning on culture, aligning on literacy, essential to anything data related.
81
00:05:22,657 --> 00:05:27,517
So you have mentioned that you're leading data governance, politician in that
82
00:05:27,517 --> 00:05:32,497
respect, but also you're in charge of the customer data strategy as well. Well,
83
00:05:32,816 --> 00:05:39,256
How pervasive is AI? How much pressure is there within the bank in terms of
84
00:05:39,256 --> 00:05:41,516
the things you're trying to do with AI at the moment?
85
00:05:41,576 --> 00:05:45,336
Is there a ton of pressure around it or is it very measured because of the high
86
00:05:45,336 --> 00:05:48,016
regulation being part of a large financial institution?
87
00:05:48,256 --> 00:05:53,076
Yeah, I'd say the latter. First Citizens has a history of doing everything in a measured way.
88
00:05:53,236 --> 00:05:57,596
And we are generally not the first to market in these spaces.
89
00:05:57,836 --> 00:06:01,676
I think it's good to sort of observe what's happening and where things are heading.
90
00:06:01,836 --> 00:06:05,516
Do you remember a few years ago, Sandy, like people were all about blockchain,
91
00:06:05,656 --> 00:06:08,936
right? The way that we talk about Gen AI today, it was blockchain.
92
00:06:09,096 --> 00:06:14,676
And everything was going to be, there's gonna be some component of blockchain in everything.
93
00:06:15,156 --> 00:06:20,716
And like, where are we today with that, right? So definitely a very measured approach.
94
00:06:20,896 --> 00:06:25,776
I do think any institution would be foolish not to at least do some exploratory
95
00:06:25,776 --> 00:06:32,516
work in this space. My personal opinion is that this is definitely more real than blockchain.
96
00:06:32,776 --> 00:06:36,536
And I do think this is going to be a game changer. We're still very much in
97
00:06:36,536 --> 00:06:41,316
our infancy, but I do believe that this is not going to go away the way that other things have.
98
00:06:41,516 --> 00:06:47,236
Yeah. I believe that Gen AI has put a highlight on AI just in general,
99
00:06:47,376 --> 00:06:49,616
right? So even the things have been around for a while, right?
100
00:06:49,696 --> 00:06:53,216
Everybody was hot on machine learning over the last five years.
101
00:06:53,216 --> 00:06:59,596
And now it's even accelerated because a lot of organizational executives are
102
00:06:59,596 --> 00:07:03,296
coming to individuals saying, what can I do that's different and novel?
103
00:07:03,636 --> 00:07:06,476
You know, starting with Gen AI as the point of the conversation,
104
00:07:06,676 --> 00:07:10,736
but then that evolves into other topics and areas that AI can help with,
105
00:07:10,876 --> 00:07:13,356
which then brings along a lot of machine learning opportunities.
106
00:07:13,356 --> 00:07:19,176
I, for one, find that there's still this challenge of,
107
00:07:19,396 --> 00:07:23,836
yep, these are all great ideas and the technology can move quickly,
108
00:07:23,916 --> 00:07:29,696
but most places still struggle getting their house in order in terms of ensuring
109
00:07:29,696 --> 00:07:31,376
they have the right data and data quality.
110
00:07:31,943 --> 00:07:37,463
Yeah, listen, I think that there are a few fundamental things like the term
111
00:07:37,463 --> 00:07:40,583
Gen AI and the work around it sounds so exciting.
112
00:07:41,203 --> 00:07:44,283
It's hard not to get sort of get wrapped up in it.
113
00:07:44,363 --> 00:07:48,063
But this work, these initiatives, these platforms and these technologies are
114
00:07:48,063 --> 00:07:51,843
still dependent on the things that we've been working on for the last decade.
115
00:07:52,203 --> 00:07:57,023
Good data governance, good data quality, excellent metadata strategy.
116
00:07:57,163 --> 00:07:59,083
Like these are still foundational.
117
00:07:59,463 --> 00:08:05,623
And I think that these are essential. Poor data quality will skew results, will create bias.
118
00:08:05,923 --> 00:08:11,163
You know, it's interesting how we're so ready to try this and look at ways that
119
00:08:11,163 --> 00:08:16,843
we can leverage it without really understanding the criteria like needed for the foundation.
120
00:08:17,003 --> 00:08:20,743
Like, you know, what does a good model look like? How do we know when we're done training a model?
121
00:08:20,923 --> 00:08:25,423
That's a very basic question that I don't think I've heard a good answer for.
122
00:08:25,663 --> 00:08:31,383
You know, I would say we know that the content that's produced out of Gen EI isn't 100% accurate.
123
00:08:31,643 --> 00:08:35,663
And so the question is, how accurate is it? Is another question that we can't
124
00:08:35,663 --> 00:08:37,663
really quantify very easily.
125
00:08:37,843 --> 00:08:42,123
We're still in that phase of asking questions that we still don't know the answer to.
126
00:08:42,303 --> 00:08:46,203
I'd almost be astonished if people felt strong that they knew the answers to
127
00:08:46,203 --> 00:08:48,143
every question in this space.
128
00:08:48,403 --> 00:08:53,323
What I'm finding to be the case is that even if you think you're getting it
129
00:08:53,323 --> 00:08:58,483
part of the answer, the answer changes just down the line because there's a new innovation.
130
00:08:58,823 --> 00:09:04,443
There's a new way of handling it. I went to Snowflake's event last week and
131
00:09:04,443 --> 00:09:09,503
they were showing their front end to JetAI capabilities for many different models.
132
00:09:09,723 --> 00:09:14,003
And a lot of the use cases they showcased, it had citations.
133
00:09:14,003 --> 00:09:17,963
You can click on the citation and it will show you the document it found the
134
00:09:17,963 --> 00:09:21,283
answer from and the section of the document and highlight it for you.
135
00:09:21,383 --> 00:09:26,603
So So in the past, people said, well, I can't cite what my model is coming from.
136
00:09:26,683 --> 00:09:28,463
And they showed that you can.
137
00:09:28,583 --> 00:09:32,183
You can do that. There's opportunity for that kind of engagement with Gen AI,
138
00:09:32,363 --> 00:09:36,103
which is, again, moving the needle forward and answering a question that people
139
00:09:36,103 --> 00:09:37,363
couldn't answer before, right? Yeah.
140
00:09:37,687 --> 00:09:41,507
It's definitely moving in that direction. And I was super excited just to see
141
00:09:41,507 --> 00:09:44,467
that one capability that I hadn't seen anywhere else.
142
00:09:44,607 --> 00:09:48,467
That's an interesting point is like, how do you create a roadmap, right?
143
00:09:48,527 --> 00:09:52,767
Even if it's like a couple quarters, right? And say, okay, I'm going to go do something.
144
00:09:52,907 --> 00:09:55,827
And you'd be foolish to think that you'd ideate on something,
145
00:09:55,967 --> 00:09:58,887
create a pilot, define an outcome, understand the dependencies,
146
00:09:59,047 --> 00:10:01,867
and then go do it even within a year.
147
00:10:01,967 --> 00:10:06,067
Like that's a thing kind of aggressive. But if the pace of change is faster
148
00:10:06,067 --> 00:10:09,607
than that, then you're almost always playing from behind.
149
00:10:09,767 --> 00:10:11,787
You know, I'd say it's sort of tangentially related to this.
150
00:10:11,827 --> 00:10:15,747
I was talking to somebody about data, data programs, and how I've been in and
151
00:10:15,747 --> 00:10:18,767
out of so many institutions, both as an FTE and a consultant.
152
00:10:18,947 --> 00:10:23,067
I've never implemented anything the same way twice, whether it was like rolling
153
00:10:23,067 --> 00:10:28,007
out a data catalog or putting in data quality framework or even setting up governing bodies.
154
00:10:28,007 --> 00:10:32,367
And I always caution others, if they feel like they have a data governance playbook,
155
00:10:32,567 --> 00:10:34,787
there are thematic things that you will follow.
156
00:10:34,907 --> 00:10:38,327
And there are themes and principles that will guide you. But you will never
157
00:10:38,327 --> 00:10:40,547
do something the same way twice, period.
158
00:10:40,827 --> 00:10:45,127
You shouldn't, right? Every institution will have its own sort of nuanced way
159
00:10:45,127 --> 00:10:47,927
of working. And Gen AI just is an exponential.
160
00:10:48,247 --> 00:10:51,527
And now you layer in technology and the pace of change in technology.
161
00:10:52,007 --> 00:10:55,347
And I told this person that I was looking at work that I did three and a half
162
00:10:55,347 --> 00:10:58,867
years ago. I thought I can go back and leverage some of that work to what we're doing.
163
00:10:58,987 --> 00:11:01,767
In many cases, that work is obsolete today.
164
00:11:01,987 --> 00:11:06,007
The way of thinking about data, even three and a half years ago,
165
00:11:06,107 --> 00:11:09,067
is so different today and has become obsolete.
166
00:11:09,267 --> 00:11:14,087
And so that's another question that's come up, which is if we endeavor to do
167
00:11:14,087 --> 00:11:17,927
something and we create a plan to do it and it takes 12 months,
168
00:11:18,127 --> 00:11:19,847
how do we adjust for the change?
169
00:11:20,107 --> 00:11:23,067
How do we be good about seeing the next corner in the road?
170
00:11:23,287 --> 00:11:28,727
Yeah. Well, one, it makes me laugh that we're talking about playbooks and really
171
00:11:28,727 --> 00:11:32,047
not having a playbook because I really think of it more as a checklist.
172
00:11:32,347 --> 00:11:35,167
What are the handful of things we need to get right?
173
00:11:35,716 --> 00:11:40,976
And, you know, we need to adjust those criteria to meet the culture of the organization,
174
00:11:41,216 --> 00:11:42,536
where they are in their maturity.
175
00:11:42,756 --> 00:11:47,276
But I'm curious, Junaid, because you talked about multiple acquisitions over the last few years.
176
00:11:47,496 --> 00:11:53,536
So as you're onboarding your acquired bank, I guess, are you changing or really
177
00:11:53,536 --> 00:11:58,616
reflecting differences in the banks and the cultures and their expectations
178
00:11:58,616 --> 00:12:00,736
during those integration points?
179
00:12:00,736 --> 00:12:05,756
Listen, I think there's, again, like lots to be said about culture and the cultural challenges.
180
00:12:05,976 --> 00:12:10,076
You know, when we think about integrating data, our biggest challenges in simulating
181
00:12:10,076 --> 00:12:14,936
the data is that there are just different approaches that the different institutions have taken.
182
00:12:15,616 --> 00:12:20,696
Hi, Sandy here. Pardon the interruption. Junaid is about to introduce the concept
183
00:12:20,696 --> 00:12:25,896
of a CDE, and I wanted to make sure everybody was on the same page in terms of what that is.
184
00:12:26,016 --> 00:12:28,276
So a CDE is a critical data element.
185
00:12:28,576 --> 00:12:32,196
So for example, in a banking context like First Citizens Bank,
186
00:12:32,416 --> 00:12:35,316
a CDE might include something like customer account numbers,
187
00:12:35,616 --> 00:12:38,436
transactional amounts, regulatory reporting fields.
188
00:12:38,576 --> 00:12:43,936
Now, these elements are critical because errors or inconsistencies in these
189
00:12:43,936 --> 00:12:48,796
types of elements can lead to significant operation issues, compliant risks,
190
00:12:49,056 --> 00:12:50,836
or inaccurate insights.
191
00:12:51,156 --> 00:12:53,656
You can imagine what that could lead to in the context of a bank.
192
00:12:53,796 --> 00:12:57,176
So with that in mind, let's get back to our conversation with Junaid.
193
00:12:57,627 --> 00:13:01,207
We have this notion of CDEs. That's a thematic thing that most people have.
194
00:13:01,287 --> 00:13:04,587
What is the most critical data that we want to govern?
195
00:13:04,807 --> 00:13:08,567
We have an additional layer of nuance, which is let's tier our CDE.
196
00:13:08,667 --> 00:13:12,407
So you might have a CDE that is derived from other CDEs.
197
00:13:12,587 --> 00:13:16,187
And now what is the difference between a tier one and a tier two?
198
00:13:16,367 --> 00:13:22,607
Getting alignment on these things is sometimes very difficult because you don't know when to stop.
199
00:13:22,747 --> 00:13:26,767
We had one particular case where we had a CDE where if you followed it to the
200
00:13:26,767 --> 00:13:29,147
end, it had something like 53 or 63.
201
00:13:29,427 --> 00:13:35,507
And we're not going to increase our governance from one CDE to 63 times that amount, right?
202
00:13:35,547 --> 00:13:38,807
It's just not. And so the question and the debate becomes, what's the right
203
00:13:38,807 --> 00:13:41,267
number? Where do we stop? And how do we tier?
204
00:13:41,527 --> 00:13:44,127
And things like that. So it's the nuance where we find our challenges.
205
00:13:44,467 --> 00:13:47,807
Yeah. Just in talking to organizations about their critical data,
206
00:13:48,027 --> 00:13:52,327
I think one of the first challenges that we run into is really aligning on a
207
00:13:52,327 --> 00:13:54,467
definition of what critical truly means.
208
00:13:54,627 --> 00:14:00,247
Yeah. And then from there, starting to align data that fits that criteria for criticality.
209
00:14:00,367 --> 00:14:03,987
Yeah, it's like, again, it's like we know that thematically we should identify
210
00:14:03,987 --> 00:14:08,887
critical data and then the definition of criticality will vary from organization to organization.
211
00:14:09,187 --> 00:14:14,587
And what's critical to someone in reg reporting will be very different to someone
212
00:14:14,587 --> 00:14:16,047
who sits in marketing. today.
213
00:14:16,107 --> 00:14:21,167
Yeah, I love the comment you made earlier where you're not going to take a playbook
214
00:14:21,167 --> 00:14:25,827
and repeat it in the next org or even division of an organization, right?
215
00:14:25,947 --> 00:14:30,547
And apply it because it is unique for the reasons Anjali stated and then some.
216
00:14:30,587 --> 00:14:34,607
Every organization even is structured differently and you need to restructure
217
00:14:34,607 --> 00:14:37,967
how you do governance because the org structure is inherently different.
218
00:14:38,147 --> 00:14:40,807
They may have business and IT, they may have something in the middle.
219
00:14:40,867 --> 00:14:43,647
They may have a product team, an engineering team and a business team,
220
00:14:43,727 --> 00:14:45,287
right? There's no one size.
221
00:14:45,627 --> 00:14:49,247
Yeah, there's absolutely no one size fit, but you have to find that middle ground
222
00:14:49,247 --> 00:14:53,187
of having that checklist, having maybe some frameworks that you can reuse.
223
00:14:53,547 --> 00:14:57,387
You know, parts of to ensure that you're covering all the bases, right?
224
00:14:57,567 --> 00:15:00,787
At least that's the way I think about it. So, and I also agree with you that
225
00:15:00,787 --> 00:15:03,207
things are outdated as soon as you put them on paper.
226
00:15:03,387 --> 00:15:08,707
One thing that I believe is completely outdated is how we think about data quality
227
00:15:08,707 --> 00:15:11,067
within organizations and how we address it.
228
00:15:11,167 --> 00:15:15,007
So I wanted to maybe jump on that topic with you a little bit.
229
00:15:15,107 --> 00:15:19,147
Given that you're in a highly regulated environment, you're probably doing things
230
00:15:19,147 --> 00:15:23,727
a little bit of an old school way, is my guess. There's probably some legacy systems involved.
231
00:15:24,007 --> 00:15:28,747
If the organization were to move on to AI, what do you think would be critical
232
00:15:28,747 --> 00:15:30,407
to get right before you start that?
233
00:15:30,627 --> 00:15:36,807
I think that there's, without a doubt, out, there are probably three disciplines
234
00:15:36,807 --> 00:15:39,467
in data that I would say are essential.
235
00:15:39,647 --> 00:15:42,227
And again, this is debatable. You could add four or five, right?
236
00:15:42,927 --> 00:15:47,987
And I would say your metadata strategy, like your underlying metadata is crucial
237
00:15:47,987 --> 00:15:51,927
in having a really well thought out, robust.
238
00:15:52,622 --> 00:15:58,082
Data catalog. Without a doubt, good data quality. You need data that is fit for purpose.
239
00:15:58,322 --> 00:16:04,062
And you need to be really clear on what the purpose is and what the criteria of good looks like.
240
00:16:04,182 --> 00:16:06,802
So data quality is another discipline that I would call essential.
241
00:16:07,002 --> 00:16:11,062
And then your governance and oversight function, the policies and procedures
242
00:16:11,062 --> 00:16:15,142
that define how you do your metadata strategy and your data quality,
243
00:16:15,342 --> 00:16:18,442
the usage of things like AI. These are three essential.
244
00:16:18,902 --> 00:16:22,542
Without a doubt, if you're not thinking about these three, you would run into
245
00:16:22,542 --> 00:16:25,842
trouble with any initiative, especially a Gen AI initiative.
246
00:16:26,142 --> 00:16:30,722
And then when I think about other fundamentals to a Gen AI initiative,
247
00:16:31,002 --> 00:16:35,202
there are probably three things that I would tell people to be comfortable with.
248
00:16:35,282 --> 00:16:37,182
Machine learning and data science.
249
00:16:37,402 --> 00:16:41,882
You should be grounded in the basics and fundamentals in both of these.
250
00:16:42,022 --> 00:16:43,882
I would advise people to learn programming.
251
00:16:44,022 --> 00:16:47,822
Python, R, and these languages Languages are very easy to learn.
252
00:16:47,962 --> 00:16:53,502
I grew up, you know, I'm going to date myself a little bit. Like we studied C and C++ in college.
253
00:16:53,762 --> 00:16:59,002
And the amount of debugging for the older list, all the, you know, commiserating with me.
254
00:16:59,082 --> 00:17:03,422
These new languages you can honestly learn in a weekend. It's not very, very hard.
255
00:17:03,602 --> 00:17:08,522
And then the other sort of concept or skill that I tell people to develop is
256
00:17:08,522 --> 00:17:09,942
be comfortable with math.
257
00:17:10,122 --> 00:17:16,562
That's a little bit harder for some. But these algorithms rely on some foundational
258
00:17:16,562 --> 00:17:21,302
understanding of probability, statistics, some linear algebra.
259
00:17:21,582 --> 00:17:24,382
I think doing those data disciplines is foundational.
260
00:17:24,722 --> 00:17:29,082
And then really thinking about these other things and grounding yourself in
261
00:17:29,082 --> 00:17:31,462
some basic knowledge is essential.
262
00:17:31,722 --> 00:17:35,342
We talk about Gen AI without talking about all those other things.
263
00:17:35,542 --> 00:17:39,522
Grounding yourself and experience in those is essential. And depending on your
264
00:17:39,522 --> 00:17:43,242
level, a certain amount of effort or granularity you'd want to get into with those.
265
00:17:43,362 --> 00:17:46,842
But even executives like senior leaders, I would say the same thing to ground
266
00:17:46,842 --> 00:17:50,902
yourself in what it takes to write a Python script, how to leverage those libraries,
267
00:17:51,182 --> 00:17:53,902
have an understanding of probability and statistics.
268
00:17:54,122 --> 00:17:57,622
Those, I think, are the foundational sort of concepts.
269
00:17:57,942 --> 00:18:01,402
I'm sure others would have so many other things, but those are the things that
270
00:18:01,402 --> 00:18:04,122
come to my mind. Yeah, I would agree with that list.
271
00:18:04,242 --> 00:18:06,842
I also learned C++ in college.
272
00:18:07,002 --> 00:18:11,502
And after that class, I never looked at it again. But I'm jealous of the younger
273
00:18:11,502 --> 00:18:17,562
generation because the amount of hours I probably spent looking for that road semicolon.
274
00:18:18,522 --> 00:18:24,542
And right now you can go into Hugging Face and just copy or put it in there.
275
00:18:24,602 --> 00:18:27,102
And it'll be like, oh, you're missing this here. And you're not.
276
00:18:28,442 --> 00:18:33,162
Yeah, I still have PTSD whenever I have to use a semicolon in a regular session.
277
00:18:33,682 --> 00:18:38,682
It always was worse, right? You spend like literally 15, 20 minutes looking for it. Oh, yeah.
278
00:18:38,782 --> 00:18:42,622
And then your colleague will find it in two seconds, right? It's just like, can you look at this?
279
00:18:42,762 --> 00:18:45,562
And like two seconds later, they're pointing right at the line.
280
00:18:45,662 --> 00:18:49,662
The new generation definitely has it so much better.
281
00:18:50,318 --> 00:18:56,138
The amount of tools and technology and platforms that they have at their disposal
282
00:18:56,138 --> 00:18:58,058
and free, the best part of it.
283
00:18:58,138 --> 00:19:04,498
The amount of technology at our disposal today is just mind-blowing.
284
00:19:04,738 --> 00:19:07,978
I also have another element to add to your list.
285
00:19:08,178 --> 00:19:12,398
You mentioned executives being familiar with coding languages to some extent,
286
00:19:12,478 --> 00:19:16,818
just to understand the efforts that it takes for a team to facilitate the solution of a problem.
287
00:19:16,938 --> 00:19:20,258
I think that's really important. I would flip that on its head as well with
288
00:19:20,258 --> 00:19:24,558
the teams who are developing, because oftentimes those teams are making decisions
289
00:19:24,558 --> 00:19:28,958
in terms of what questions to ask of the data to find answers.
290
00:19:29,078 --> 00:19:33,138
And they're lacking those economic concepts that they need to be aware of to
291
00:19:33,138 --> 00:19:37,218
ensure they're asking the right questions in their analysis in order to find
292
00:19:37,218 --> 00:19:39,398
the right statistical model to go after.
293
00:19:39,658 --> 00:19:42,998
Just because you have an answer to a question, it doesn't mean you're asking
294
00:19:42,998 --> 00:19:45,898
the right question. That's a very, very good point.
295
00:19:46,098 --> 00:19:51,558
And you're right. I think that there is still so much learning to be done at every level.
296
00:19:51,758 --> 00:19:56,398
The people who are actually working hands-on keys probably have the biggest challenge.
297
00:19:56,718 --> 00:20:01,318
And I would say for any data person, one of the challenges that I have isn't
298
00:20:01,318 --> 00:20:04,358
those guys, isn't the team that has hands-on keys.
299
00:20:04,678 --> 00:20:07,938
I feel like they're invested and on the journey. right?
300
00:20:08,398 --> 00:20:13,478
It's the executive sponsorship, to be honest. And I feel like getting executives
301
00:20:13,478 --> 00:20:19,858
and leaders who don't have a history in data on board is to me is an age old problem.
302
00:20:20,038 --> 00:20:23,798
And I heard somebody say like the way that executives today have to have some
303
00:20:23,798 --> 00:20:27,578
foundational understanding of finance, like they have to know earnings per share.
304
00:20:27,718 --> 00:20:32,738
They have to know EBITDA. There are many of these CEOs, they're all grounded on certain mechanics.
305
00:20:33,058 --> 00:20:35,018
I heard somebody say that they're They're going to have to equally,
306
00:20:35,038 --> 00:20:40,358
to be successful CEOs, they're going to have to be equally grounded in the information
307
00:20:40,358 --> 00:20:41,858
age when it comes to data literacy.
308
00:20:42,178 --> 00:20:45,678
Yeah, I would agree with that. I would agree with that. I was in board readiness
309
00:20:45,678 --> 00:20:50,038
events and there were a number of current board of directors,
310
00:20:50,338 --> 00:20:55,838
advisory boards, individuals that they're currently on multiple wards and some public boards.
311
00:20:55,838 --> 00:20:59,698
And one of the things that they highlighted, which I could not believe,
312
00:20:59,858 --> 00:21:05,598
was the fact that most board members are actually illiterate in terms of digital
313
00:21:05,598 --> 00:21:07,738
and data and technology concepts.
314
00:21:08,395 --> 00:21:11,355
They're all over cyber because they have to be, right? That's part of risk.
315
00:21:11,535 --> 00:21:15,635
But they're unable to think about, is the organization doing the right things
316
00:21:15,635 --> 00:21:20,295
in order to ensure they're ahead of the curve, getting access to data or ensuring
317
00:21:20,295 --> 00:21:23,555
they're getting value from their data? They're just not there yet.
318
00:21:23,715 --> 00:21:27,515
And further, the CEO, they may delegate out to a CIO.
319
00:21:27,655 --> 00:21:29,635
But how do you find the right C-level
320
00:21:29,635 --> 00:21:33,895
executives for your team if you're unaware of what good looks like?
321
00:21:34,035 --> 00:21:37,755
Yeah, you know, I'm hopeful that that change is coming and it's coming quickly.
322
00:21:37,755 --> 00:21:40,555
The CDO office didn't exist, right?
323
00:21:40,695 --> 00:21:47,135
Not in the way it does even 10 years ago, 15 years ago. It was virtually non-existent 15 years ago.
324
00:21:47,295 --> 00:21:50,555
And it really started catching on probably 10 years ago.
325
00:21:50,635 --> 00:21:53,115
In the last five years is where you've seen probably a spike.
326
00:21:53,195 --> 00:21:56,835
I haven't seen any statistics on this, but I'd imagine it'd be like a hockey
327
00:21:56,835 --> 00:22:01,815
stick curve where in the last 10 years, you'd see this sort of like low number of CDOs.
328
00:22:01,915 --> 00:22:03,275
And so that's probably the first
329
00:22:03,275 --> 00:22:07,855
phase of this, which is the advent of the office and the role of CDO.
330
00:22:07,975 --> 00:22:12,735
And then I think what I've seen is that there's been often a misalignment of where the CDO sits.
331
00:22:13,055 --> 00:22:16,455
And I think that's still a debate. We probably have another podcast on.
332
00:22:16,635 --> 00:22:19,795
And so that's a whole other thing. So at least they exist.
333
00:22:20,095 --> 00:22:23,675
And then I think the other challenge has been, what does the CDO do?
334
00:22:23,875 --> 00:22:29,415
The CDO role is on average 24 months, some 22, 26 months, right?
335
00:22:29,415 --> 00:22:33,215
And so just as you're getting settled, there's a shift.
336
00:22:33,395 --> 00:22:37,235
There's a lot of that that comes into place. I have the personal view that a
337
00:22:37,235 --> 00:22:39,035
CDO should be a direct to a CEO.
338
00:22:39,495 --> 00:22:45,335
That's often not the case. If you can bring that CDO closer to the CEO and peers
339
00:22:45,335 --> 00:22:48,915
with everybody else, that's how you would accelerate the data literacy objective
340
00:22:48,915 --> 00:22:51,055
that exists at the C-suite. Right.
341
00:22:51,175 --> 00:22:56,175
I agree with you wholeheartedly on that. I like to think of it as you're either
342
00:22:56,175 --> 00:22:59,495
a capital C DO or you're not.
343
00:22:59,575 --> 00:23:02,955
And the capital C is when you're reporting to the CEO.
344
00:23:03,669 --> 00:23:08,269
And you have a seat at the table. You know, non-capital C is when you're in
345
00:23:08,269 --> 00:23:13,169
a division or you're part of a business unit or you're, because they have those as well.
346
00:23:13,289 --> 00:23:20,309
I've seen a lot of organizations where I have the CDO of this part of our region or our division.
347
00:23:20,949 --> 00:23:25,229
They really are in those positions. They don't own the technology stack.
348
00:23:25,429 --> 00:23:28,609
They don't own the engineering team. They don't sometimes don't even own the
349
00:23:28,609 --> 00:23:32,769
product side of it. And you're literally just brokering conversation and ideas
350
00:23:32,769 --> 00:23:36,889
for strategy without any of the ability to actually execute on it.
351
00:23:36,949 --> 00:23:40,569
That's a good point. And like, it begs the question, where should AI initiatives sit?
352
00:23:40,709 --> 00:23:45,849
Should they sit with a CDO or should they sit with a CIO or a CTO, right?
353
00:23:45,969 --> 00:23:50,609
Because it's actually a technology, but it's producing, Gen AI is a content
354
00:23:50,609 --> 00:23:54,029
producer, so it's producing data and it requires data.
355
00:23:54,149 --> 00:23:57,769
I've started seeing cheap AI officer roles too, right? Right.
356
00:23:57,769 --> 00:23:59,929
And maybe that's the answer. I'm not quite sure.
357
00:24:00,149 --> 00:24:02,429
It's another one of those questions where I don't know that there's a clear
358
00:24:02,429 --> 00:24:04,569
answer. Yeah, I could debate that all day.
359
00:24:04,789 --> 00:24:10,909
I think it just depends, right? It depends on how the organization defines those two roles.
360
00:24:11,229 --> 00:24:17,429
And if both roles are in charge of leveraging data to push business strategy
361
00:24:17,429 --> 00:24:20,489
in the right direction, it's the same role.
362
00:24:20,809 --> 00:24:25,669
I think that's a valid argument. That's the only way I look at it. It's the same role to me.
363
00:24:25,789 --> 00:24:29,769
You basically undress your chief data officer, the moment you hire a chief AI
364
00:24:29,769 --> 00:24:34,629
officer, because the AI officer is going to create that value within the workstream,
365
00:24:34,729 --> 00:24:37,609
within the operations, within the business.
366
00:24:37,949 --> 00:24:41,549
And a chief data officer then becomes a governance body immediately.
367
00:24:41,949 --> 00:24:45,129
To me, it just depends on what you're looking to achieve with that role.
368
00:24:45,369 --> 00:24:49,909
And maybe that's the evolution, the CDAO title, where the A is then replaced,
369
00:24:50,129 --> 00:24:54,989
where the analytics is then replaced with AI, perhaps, is where we're headed.
370
00:24:55,169 --> 00:24:58,369
I think it's going to be interesting to look back at this conversation in 24
371
00:24:58,369 --> 00:25:02,089
months and see where the answers are, right?
372
00:25:02,209 --> 00:25:04,289
And say, where is the industry headed?
373
00:25:04,529 --> 00:25:08,329
That's one of the reasons I like talking about this topic, about how just new
374
00:25:08,329 --> 00:25:12,769
it is and how we don't know the answers. I think it's one of the most exciting aspects of it.
375
00:25:12,889 --> 00:25:17,209
Yeah, well, it might be actually interesting to reflect back on this conversation.
376
00:25:17,369 --> 00:25:22,009
In about two months, I have an upcoming talk with a friend and a thought leader
377
00:25:22,009 --> 00:25:27,589
in data and analytics in July, where we talk about embedding data and analytics
378
00:25:27,589 --> 00:25:29,969
within an organization's data culture.
379
00:25:30,129 --> 00:25:34,389
And really it comes down to the responsibility of the chief data officer.
380
00:25:35,009 --> 00:25:40,149
Yet we haven't really seen our CDOs being as effective as we hope.
381
00:25:40,229 --> 00:25:43,189
And now where does AI actually sit?
382
00:25:43,349 --> 00:25:49,369
And our intention is to actually start solutioning some of the challenges that we've seen,
383
00:25:49,815 --> 00:25:55,595
with a group of CDOs as part of this discussion. I'll share back what we learn later in the summer.
384
00:25:55,775 --> 00:25:57,995
That would be interesting. You know, you got me thinking about something,
385
00:25:58,195 --> 00:26:02,595
Anjali. So, you know, Sandy and I talked about data democratization a few months ago now, I think.
386
00:26:02,675 --> 00:26:05,975
You just made me think of the concept of AI democratization.
387
00:26:06,395 --> 00:26:09,835
You know, I'm a big proponent to data democratization. I think,
388
00:26:09,835 --> 00:26:13,975
you know, there are noted challenges around governance and security and access.
389
00:26:14,315 --> 00:26:17,375
And somebody asked us the question on our panel, other than data governance,
390
00:26:17,375 --> 00:26:21,375
what's the best way to ensure appropriate use of data and data democratization?
391
00:26:21,495 --> 00:26:23,875
And I said, I know you don't want to hear that the answer is data governance,
392
00:26:23,975 --> 00:26:26,575
but the answer is data governance is what I said.
393
00:26:26,695 --> 00:26:30,975
And I think that you want to put data out there. I'm of the belief where we
394
00:26:30,975 --> 00:26:33,775
should view any institution where everybody's a data person.
395
00:26:33,935 --> 00:26:36,515
If you're an accountant and you're producing data, consuming data,
396
00:26:36,635 --> 00:26:39,315
and you're producing an earnings per share report, you're a data person.
397
00:26:39,595 --> 00:26:42,735
And so I have the belief that you should democratize data. Now the question
398
00:26:42,735 --> 00:26:44,535
is, do we democratize AI?
399
00:26:44,775 --> 00:26:49,275
And do we put AI in the hands of everybody? I'm convinced it's on the doorstep.
400
00:26:49,675 --> 00:26:52,995
At the end of the day, if you look at enterprise applications.
401
00:26:53,975 --> 00:26:57,775
They're every single enterprise application company in the world.
402
00:26:57,895 --> 00:27:02,015
There is no enterprise application company that is in the market today.
403
00:27:02,175 --> 00:27:05,155
And if unfortunately they're not doing this, they won't be in the market for long.
404
00:27:05,275 --> 00:27:08,935
But they are all infusing Gen AI capabilities into their platforms.
405
00:27:08,935 --> 00:27:14,955
I haven't seen a demo in the last six months that does not include some aspect of that.
406
00:27:15,155 --> 00:27:19,435
And if you work for a company and they're not doing this, find another job.
407
00:27:19,515 --> 00:27:22,375
I'm telling you this right now, because they're going to be left behind.
408
00:27:22,555 --> 00:27:26,755
Everyone has it. And if I think about AI governance within an enterprise,
409
00:27:27,155 --> 00:27:32,455
one of the tasks that they have is understanding how is that Gen AI used within
410
00:27:32,455 --> 00:27:36,955
an enterprise application that we now license or looking to license and ensuring
411
00:27:36,955 --> 00:27:39,695
that it's not going to lead us down a bad path, right?
412
00:27:39,855 --> 00:27:43,675
I mean, if you look at just some of the things that came up with HR functions,
413
00:27:43,915 --> 00:27:45,135
for example, that was an issue.
414
00:27:45,255 --> 00:27:49,035
Group inside, it was an issue on performance management, helping people figure
415
00:27:49,035 --> 00:27:50,435
out who to promote, those kinds of things.
416
00:27:50,515 --> 00:27:54,755
They were all issue areas that I think have evolved since those issues came
417
00:27:54,755 --> 00:27:56,915
up, I don't know, it was a year ago or so.
418
00:27:57,055 --> 00:28:00,815
But it just kind of shows every enterprise application is moving in that direction
419
00:28:00,815 --> 00:28:04,675
because they want to make the use of the application as frictionless as possible
420
00:28:04,675 --> 00:28:06,015
and user-friendly as possible.
421
00:28:06,135 --> 00:28:09,835
And there's nothing more user-friendly than a chatbot. So that's where they're headed.
422
00:28:10,035 --> 00:28:15,135
I mean, even Copilot on Microsoft, they're throwing it onto your desktop to
423
00:28:15,135 --> 00:28:17,775
the point where I don't have to go navigate to a file.
424
00:28:18,235 --> 00:28:22,475
I could just say, give me X, Y, Z, and a file opens up on my Microsoft computer,
425
00:28:22,675 --> 00:28:23,815
right? So I have an Apple.
426
00:28:23,935 --> 00:28:26,415
But anyway, you get my point? I think that's where we're headed.
427
00:28:26,655 --> 00:28:30,775
I also have seen, we're going to have another episode where we talk about the
428
00:28:30,775 --> 00:28:34,935
Snowflake conference, but I'll just add this to quick little thing I saw at
429
00:28:34,935 --> 00:28:37,695
the conference, which was they were releasing all these capabilities.
430
00:28:38,095 --> 00:28:41,375
And at one point they said, everyone should be able to build a chatbot.
431
00:28:41,535 --> 00:28:44,935
And then he kept repeating it. And then he said, we're going to grab an audience member.
432
00:28:45,035 --> 00:28:48,035
We're going to have them come up and build a chatbot. And in five minutes,
433
00:28:48,175 --> 00:28:52,655
this person who isn't a coder, and I believed her, she was in the back of the
434
00:28:52,655 --> 00:28:56,055
in the middle of the room, they called her name, she came up,
435
00:28:56,115 --> 00:28:59,375
she did what they told her, she hit the wrong button a few times,
436
00:28:59,535 --> 00:29:04,255
they messed up the demo for a second, but they built a chatbot bot off of PDFs.
437
00:29:04,255 --> 00:29:06,895
And it took her five minutes, and it worked.
438
00:29:07,095 --> 00:29:10,495
Pretty enticing demo. And I'm sitting there going, well, if anybody can do that,
439
00:29:10,615 --> 00:29:15,655
imagine the possibilities of the ability to just really ensure that all the
440
00:29:15,655 --> 00:29:19,315
knowledge within an organization is readily available to everybody within the organization.
441
00:29:19,535 --> 00:29:25,915
That's amazing, right? And it adds to the excitement of what's out there, what the potential is.
442
00:29:26,075 --> 00:29:28,055
It sounds amazing. And I think you're right.
443
00:29:28,295 --> 00:29:31,915
Institutions that don't embrace this will likely get left behind.
444
00:29:32,135 --> 00:29:38,035
Even if it's a measured approach, I think it'd be advisable to get grounded in the topic.
445
00:29:38,355 --> 00:29:42,275
Absolutely. I avoided it when it first happened as well. Because like you,
446
00:29:42,495 --> 00:29:45,095
we've all been burned by these ideas, right?
447
00:29:45,575 --> 00:29:50,815
When I graduated from college, the dot-com era busted out of my first job.
448
00:29:50,815 --> 00:29:53,835
I was laid off because those jobs were gone overnight.
449
00:29:53,995 --> 00:29:58,055
The market failed and all these startups turned into nothings, right?
450
00:29:58,155 --> 00:30:00,215
I was part of a unicorn company and
451
00:30:00,215 --> 00:30:04,235
that turned into, I think my stock was worth $2 by the time I was gone.
452
00:30:04,415 --> 00:30:08,275
So it felt very much like that.
453
00:30:08,435 --> 00:30:12,215
And I think we will see that wave of companies getting created,
454
00:30:12,335 --> 00:30:17,375
entering semi-unicorn status, and then being thrown out the door because maybe
455
00:30:17,375 --> 00:30:21,635
they were trying to build something that should be part of a different application, et cetera.
456
00:30:21,855 --> 00:30:24,255
So that's definitely going to happen, I think, in this wave.
457
00:30:24,415 --> 00:30:29,595
But what happened to me initially was, oh, here we go, another interesting flash in the pan thing.
458
00:30:29,795 --> 00:30:32,555
Even Gartner was like, this is a shiny object a year ago, right?
459
00:30:32,655 --> 00:30:37,335
And now we're all over it all I would say I have a slightly different view of
460
00:30:37,335 --> 00:30:40,975
what will happen to all these tech companies.
461
00:30:41,650 --> 00:30:47,970
My view is that I'm with you, that they sprout up and they create a bunch of very niche solutions.
462
00:30:48,390 --> 00:30:51,690
But what I've observed is that they won't necessarily go out of business the
463
00:30:51,690 --> 00:30:53,030
way that dot-coms did overnight.
464
00:30:53,210 --> 00:30:58,270
I see them just getting assimilated, like being acquired by other larger companies.
465
00:30:58,530 --> 00:31:00,590
And then I think that you'll see consolidation.
466
00:31:00,910 --> 00:31:04,790
Sometimes I think that's why they start these companies is to plan to get acquired.
467
00:31:05,090 --> 00:31:09,610
Right. And some players in the marketplace, some cloud providers out there,
468
00:31:09,690 --> 00:31:15,090
keep it nameless this for now, but they are playing smart in terms of enabling
469
00:31:15,090 --> 00:31:19,290
a platform that people can build on. I've seen it in a number of instances.
470
00:31:19,590 --> 00:31:23,430
Salesforce has always been one of those players where, as an example,
471
00:31:23,570 --> 00:31:27,150
and they've always kind of been in this marketplace of build for Salesforce,
472
00:31:27,410 --> 00:31:28,750
build an app on top of Salesforce.
473
00:31:29,310 --> 00:31:33,950
And anybody who has a startup, and I'm not saying Salesforce specifically,
474
00:31:34,270 --> 00:31:38,710
but they should look at the marketplace and say, should I make a bet on a provider
475
00:31:38,710 --> 00:31:43,590
on whether it's AWS, Microsoft, Salesforce, Snowflake, Databricks, whatever it may be.
476
00:31:43,690 --> 00:31:47,810
Should I make a bet on a provider and build my application to work specifically
477
00:31:47,810 --> 00:31:52,570
for that platform and already is integrated inherently into that platform?
478
00:31:52,810 --> 00:31:56,130
Because that provider may gobble you up. That's your exit strategy,
479
00:31:56,270 --> 00:31:59,250
right? I'm going to make it so compelling for them. And I'm going to sell to
480
00:31:59,250 --> 00:32:00,530
every single one of their clients.
481
00:32:00,690 --> 00:32:06,090
And I'm going to ensure that I make a splash every single year with them from
482
00:32:06,090 --> 00:32:08,410
a revenue standpoint so that they can't avoid me.
483
00:32:08,570 --> 00:32:12,130
And at some point, they're going to acquire you, especially if it's on their
484
00:32:12,130 --> 00:32:15,650
platform, it's fully integrated, it's kind of native to their ecosystem.
485
00:32:16,130 --> 00:32:19,210
I think that's an easy play for them to say, hey, I'm going to go ahead and
486
00:32:19,210 --> 00:32:21,270
acquire you and relieve you of this.
487
00:32:22,410 --> 00:32:28,030
I'm with you. What would be your advice to someone who's either mid-career,
488
00:32:28,130 --> 00:32:31,290
kind of later in their career, just trying to keep up with the madness?
489
00:32:31,830 --> 00:32:35,670
You know, one advice that I've given recently is, you know, I think that in
490
00:32:35,670 --> 00:32:39,590
the last 10 years, there's been sort of these nuanced roles that people fit into.
491
00:32:39,790 --> 00:32:45,010
You know, I'm of the belief that versatility is a key to success.
492
00:32:45,350 --> 00:32:51,330
Having this skill to work across data disciplines is the advice that I would
493
00:32:51,330 --> 00:32:54,290
give people coming up. I'll give like a basketball analogy.
494
00:32:55,001 --> 00:32:59,241
So in basketball, you have these positions, you have a point guard and you have a center, right?
495
00:32:59,341 --> 00:33:03,161
And historically very different, right? In terms of size and style of play.
496
00:33:03,341 --> 00:33:07,941
But the most exciting basketball and the ones that people, teams I see most
497
00:33:07,941 --> 00:33:11,901
successful is where there is this concept of positionless players,
498
00:33:12,141 --> 00:33:15,281
people who can play at any position at any given time.
499
00:33:15,281 --> 00:33:20,941
I think the future of like data professionals is someone who is so versatile
500
00:33:20,941 --> 00:33:26,021
and can be sort of positionless and have enough skill working across all these
501
00:33:26,021 --> 00:33:29,521
disciplines is the advice that I would give. And be innocent.
502
00:33:29,781 --> 00:33:32,241
That's the other piece of advice. Expect continuous learning.
503
00:33:32,381 --> 00:33:35,501
And if you aren't going to embrace it, this isn't for you.
504
00:33:35,621 --> 00:33:40,441
I wholeheartedly agree with you primarily because I see myself as someone who
505
00:33:40,441 --> 00:33:43,841
started in left field and ended up going through all these iterations.
506
00:33:43,841 --> 00:33:45,861
In my career was not a straight line.
507
00:33:46,021 --> 00:33:53,121
It was a bunch of loop-de-loops to the point where the way I look at it is you
508
00:33:53,121 --> 00:33:58,761
cannot influence change without understanding what is required of the other person's role.
509
00:33:59,081 --> 00:34:03,521
So how are you going to influence them to primarily do the right thing if you
510
00:34:03,521 --> 00:34:05,921
don't have enough understanding to push them in that direction,
511
00:34:06,081 --> 00:34:07,301
to get what you need done?
512
00:34:07,461 --> 00:34:10,961
I think that's going to be critical. And the only way to get there is to have
513
00:34:10,961 --> 00:34:15,541
the knowledge of of how those individuals have to work and what they're trying
514
00:34:15,541 --> 00:34:18,921
to achieve and what they're doing. Is that the right thing?
515
00:34:19,061 --> 00:34:23,261
Or is there something else and you should be pushing them in the right direction?
516
00:34:23,621 --> 00:34:27,201
Yeah, fair enough. I will say that I think I personally prefer a career that
517
00:34:27,201 --> 00:34:29,661
was like loop to loop versus a straight line as well.
518
00:34:29,741 --> 00:34:34,341
So I think you're fortunate. And I think it probably adds to your skill set.
519
00:34:34,761 --> 00:34:36,241
I'm very fortunate, yes.
520
00:34:36,741 --> 00:34:41,781
I didn't appreciate it until later in life. So we completely went off script on this.
521
00:34:42,521 --> 00:34:46,101
For those listeners out there, we were actually going to talk about how Gen
522
00:34:46,101 --> 00:34:47,641
AI can solve data quality issues.
523
00:34:47,901 --> 00:34:53,081
I still think it's an incredible topic. Yeah, I do think Gen AI is prime for
524
00:34:53,081 --> 00:34:54,941
solving data quality issues.
525
00:34:55,101 --> 00:34:57,641
You know, I'll say this, you know, I think I shared with you,
526
00:34:57,661 --> 00:35:00,761
I started my career as an accountant. My first job was...
527
00:35:01,078 --> 00:35:05,718
There's an intern reconciling accounting entries for inter-entry transactions.
528
00:35:06,298 --> 00:35:10,118
What it was, was an accounting problem, a data problem. And the data problems
529
00:35:10,118 --> 00:35:14,198
were missing data, format of data, just anomalies around data.
530
00:35:14,358 --> 00:35:19,258
And if you think about it, Gen AI is perfect for these kinds of problems to
531
00:35:19,258 --> 00:35:21,038
solve when you think about data cleansing.
532
00:35:21,238 --> 00:35:27,418
But yeah, maybe another podcast hyper-focused on data quality and Gen AI,
533
00:35:27,518 --> 00:35:28,838
data governance and Gen AI.
534
00:35:29,098 --> 00:35:33,138
Yeah, absolutely. I would love it. I went in last week with that in mind,
535
00:35:33,158 --> 00:35:37,878
and I actually ran around the conference floor looking at companies who are
536
00:35:37,878 --> 00:35:42,738
solving data quality issues with ML, AI, Gen AI capabilities.
537
00:35:43,558 --> 00:35:45,898
And they are many, many out there.
538
00:35:46,078 --> 00:35:50,238
It was eye-opening to see that because it's addressing an old problem in new
539
00:35:50,238 --> 00:35:53,898
and novel ways that removes some of the headaches of how to solve the problem.
540
00:35:54,038 --> 00:35:56,238
Definitely interesting conversation to be had.
541
00:35:56,578 --> 00:36:00,558
Thank you so much for having me. It was so wonderful just chatting without boundaries.
542
00:36:00,558 --> 00:36:04,418
Yeah, as advertised, I told you that when we first chatted about the episode,
543
00:36:04,558 --> 00:36:06,258
I said, hey, we're going to have a construct.
544
00:36:06,498 --> 00:36:10,938
But if we go in another direction, we'll just take it there if the conversation is good.
545
00:36:11,038 --> 00:36:15,318
And it was. So thank you again for your time, Junaid. And I hope you have a
546
00:36:15,318 --> 00:36:18,518
fabulous weekend. Thank you, Sandy. Thank you, Angela. Appreciate it.
547
00:36:18,640 --> 00:36:32,155
Music.