CIO Playbook for Enterprise AI | CXOTalk #810
Today on Episode #810 of CXOTalk, we're discussing enterprise AI with two of the top CIO advisors in the business. Welcome to Tim Crawford and Isaac Sacolick. Boards and CEOs expect CIOs to have a strategy, to have a defined vision in terms of where the organization can take advantage of these technologies. There's a lot of pressure top-down, but there's also a lot of bottom-up activity, and that's both people who are very excited about what they can do with AI and spending a lot of their time testing out these different tools, and those that are really fearful, scared about their jobs. What are they going to be doing as a marketer or a software developer, as prompting becomes kind of commonplace in their technology acumen? What should CIOs be doing in terms of championing these initiatives and bringing the pieces together? That's the $64,000 question that every CIO I work with is asking. Where does it fit in? Where does it not fit in? How do we start to navigate some of the things that Isaac had mentioned starting with just the innovation but not ignoring the people element to it as well because there is a lot of concern about job loss and "this is going to replace me." Most of it inappropriately thinking along those lines. I don't think it'll have as dramatic an impact as people might think. But let's
not forget; we have a lot of history that shows that there's change that comes with innovation. The real issue for CIOs is how to navigate through that to ensure that you are embracing generative AI and AI. Granted, AI has been around for decades. Generative AI, a little more recent. But at the same time, making sure that you put up the guardrails, where appropriate, and that you don't over-rotate, which I think we'll talk about what that really means. Then, at the same time, be a leader. Be a leader to your organization both up the
chain of command as well as parallel to your position and within your own organization. Help them understand and educate them on what those opportunities are today. Frankly, that'll change over time. As we get to know these
technologies more and more, it will change. We have to be adaptable. We have to be at the front end of the line and help guide that process as leaders. That's probably the single biggest opportunity for CIOs is to be at the front of the line and help provide that guidance in a balanced way. I agree with you, Tim. You think about every introduction of a new technology, analytics, or AI capability now. We're moving up stack. We're taking things that we do today. We figure
out a workflow around it, a way of working. All of a sudden, a new technology comes into play and says, "You know what? I don't need to do keyword searching anymore. I'm doing prompting. And I can continue to re-prompt," and that's changing the workflow. It's changing the skill set of people and how they're interacting with these tools. That breeds some who just lose a ton of their time because they're just experimenting and playing with the latest gadget that's come out. Others are just fearful about what this all means. "I'm a developer. I know how to code. What's Copilot going to mean for me? I
already know how to do development very well." The same thing on the marketing end in terms of building campaigns or writing content. That's what's happening in the trenches, and when you talk to what CIOs should be doing, the first thing that I think about is coming up with the types of problems that you want people to be focused on. Where is this really going to have impact in your organization? How do you register the types of value that a particular AI is bringing to an operation? How do expose that? How do you make sure that when people are using a particular tool, they're using data in a sanctioned way? They're using the right tools with the right data? That's the leadership role, bringing it all together and saying, "We're focused here. Here is our short-term vision. Here is the strategy around it. Here is the registry that we want
to use to capture ideas. And here is how you provide feedback to us about what's working." How is this different than any other technology initiative? What Isaac was just describing sounds like plain vanilla leadership for technology. We get lost in the details. We forget the big picture. What's our strategy? How is any of this different? If you look through the big tectonic shifts in technology, whether it was going from mainframes to distributed computing or the Internet or cloud, all of those kind of sat behind the scenes within IT, at least initially, and then eventually would get out to the public domain. One of the things we've seen with generative AI is it started in the public domain, and we have ChatGPT to thank for that. Now, everybody and their brother has access,
and most people today, the average person has used ChatGPT in some way, shape, or form. They've been exposed to this new innovation and seen what it potentially could do. Now, that's coming from a position of just experimentation. But that does filter into driving what enterprises then have to look at and have to think about and have to do.
Meaning, I used it to build this content. Isaac talked about marketing. Content creation is one of the three big areas that generative AI is being used for today. The content creation piece is what common folks are using generative AI for. Then they think, "Okay, well, I can do this in my personal life. How do I do it in my professional life?" to make that end of the spectrum easier to manipulate and work through. You have that coming about, which I think has just accelerated the adoption and the interest in generative AI specifically.
Some have said it's having its iPhone moment because we saw that when the iPhone came out, and there was just this mass adoption because people realized, "Oh, my gosh. I can take these 20 or 30 different functions, and it's all within one device now." Generative AI is kind of having that same moment. But we're going to learn. We're going to learn where it does fit and where it doesn't fit, as we go through the upcoming months. We're now seeing more companies looking at how to leverage large language models behind the firewall, looking at their information, starting to look at everything from textbooks in higher ed to financial information that's coming in the form of news and other content, and saying, "How do we make people smarter with this content?" and also use the capability of natural language querying and prompting so that that skillset goes down, that we don't necessarily have to have experts in every field. We have to have people who have to know how to ask the right question and being able to challenge and interpret results. I think
that's the bigger challenge here when you look at AI versus some of the other technologies. Everyone that's listening to this has heard the term "hallucinations." Generative AI will hallucinate. It'll make things up. I had someone just recently say, "Generative AI and ChatGPT can be the best liar because it builds confidence." There's actually a story that's out there now talking about an attorney that used generative AI to create a brief and then kind of double-down on that. Generative AI ended up making up the cases [laughter] and the details in the cases that didn't exist.
Is that a bad thing? Well, maybe. But maybe not because it's optimized for confidence, and so building toward confidence. But as Isaac said, there is this big question mark as to when you do a search, you can tie it back to the source. And so,
provenance of data and trust in data is really important. But with generative AI, because of the nature of how it works, there's this breaking point within the algorithm itself and how it works that we have to kind of get over and start to understand. We saw this in cloud with ephemeral workloads, workloads that come and go, instantaneously almost. And so, I think we have to go through the same type of process of learning a different way to gain confidence in results in data, and we haven't gotten there yet. We've got a ways to go. Please subscribe to our newsletter and subscribe to our YouTube channel. Just go to CXOTalk.com. We have amazing shows coming up, live shows where you can ask your questions,
and truly, genuinely, you are part of the conversation. So, check out CXOTalk.com. Where are CIOs in terms of understanding the opportunities, as well as the various risks, as well as the learning curve that Tim was just talking about? Overall, what's the state of the CIO when it comes to these issues? There's a lot of learning happening at too slow of a pace is how I would categorize this. Technology and the capabilities are moving too fast. The employees are experimenting, and the CEOs are trying to nudge the organization to be at the forefront, sensing that it's going to change business models, it's going to change customer experiences. They're hoping that it makes their workflow and their employee environments a little bit more efficient than they were in the past. CIOs really need to have a documented vision and strategy around this. I've done some writing around what goes into that: getting into a sense
of what problems to focus on, what areas of the business are going to have a material impact by experimenting with that. What's the goal around an LLM capability that maybe they're trying to build up over a horizon two and a horizon three? I think there's a lot of learning, but the risk here for CIOs is to get something out there on paper and start communicating, letting your business partners know that you are going to be the center point of putting a strategy together. There's a risk of shadow AI when a general manager says, "I need to go figure this out," and decides to go off and sign up a new platform or test a new set of capabilities. Even if it's wrong, it's going to evolve. What we are talking about now as a set of capabilities has changed since two, three months ago. Being able to do blue sky planning with
business leaders, with technologists and data scientists on a very frequent basis to say, "Is our strategy aligned or do we need to pivot or do we need to add to it?" I think that's really the goal for a CIO now is to continually do that over the course of how this technology is changing. Tim, Isaac just said that the learning that CIOs are undertaking right now is, frankly, too slow. Do you agree with that? And the important question, therefore, is what should CIOs be doing? Generative AI doesn't work in a vacuum, and the remit for the CIO is not just about generative AI. They're still navigating. CIOs are still navigating through trying to figure out how to make things like hybrid-work work, how to make remote-work work, how to think about technical debt, how to deal with cybersecurity and ransomware. All of these other big, huge, monumental issues still exist for the CIO. And so, this is just
one more thing, one really big thing that has popped up and now is on top of everything else. He's right. The learning is moving too slowly. But at the same time, this is where I think it's important to leverage your peers, leverage trusted sources. It doesn't have to be just the two of us, but it could be the two of us. It could be others. There are some really good sources out there that are actually looking at the bigger picture and starting to navigate and starting to see some of these challenges that are coming up. Then make sure that you're putting the right guardrails in place to at least bide you some time. Don't hinder innovation. Don't hinder experimentation.
I've seen everything from trying to block access to ChatGPT to blocking every .ai domain (by institutions). I don't think that's necessarily the right way to go about it. I understand why those individual companies do that. But I think it's important to find ways to expand your organization in terms of how it thinks about experimentation so it's not just the CIO learning, but it's their executive team, it's their junior folks on the team bringing fresh and new ideas to the table. It's about leveraging your organization,
and I don't think we do enough of that through the course of being an IT leader. But this is a good time to explore that and generative AI is a great place to test it out. How much emphasis right now should CIOs be placing on enterprise AI, generative AI? How much resources compared to all the other things that are on the plate? Let's make this really practical right now. I would say a lot and the reason for that
is because we know that all of these big spaces, these big boulders that we have to contend with as IT leaders, they all are going to rely more and more on data. And so, when you think about data and you think about the mounds of data that you have access to today, let alone what you have access to externally, you have to find a better way, a smarter way to be able to gain insights from that data. This is where things like generative AI can really help move that forward. We saw that recently with networking companies that are starting to move natural language processing into the process. Isaac mentioned Copilot and code development. But again, let's balance that, too. Code development is a great example of ensuring that you're not leaking intellectual property or you're not inadvertently bringing someone else's intellectual property into your code base. Those are examples where you have to find ways
to leverage this in a myriad of different ways. I think it's a total game changer for the enterprise across the suite of pillars within the company. But you have to make sure that you're using it in the right way so that you don't end up leaking confidential information or intellectual property or, vice versa, bringing that in inadvertently. CIOs are looking at their business strategies,
looking at where they are in the technical modernization, where they are in terms of security. There are a lot of companies out there that are still catching up. When you're still catching up, you look at a shiny object like AI, and your data isn't in order, you don't have a way of communicating with your employees in terms of what experiments they should be working on what experiments they shouldn't be working on. You have known security issues. I look at the entire corpus of what is the CIO driving at and say, "Okay, where are the greatest opportunities and risks?" For some industries and some companies, AI is going to be in the top three right now, and it should be in the top three. For a lot of companies, the CIO's job is to
articulate back to boards and back into their executive group, "Look. It should be in the top three. But we have no data governance strategy. We're still struggling with hybrid work," some of the examples that you shared there. We need to catch up on that, I think, regardless of where you are. I think when we talk about CIOs being too slow, I think the real issue
here is the CIOs need to come up with a point of view and a documented plan and strategy around it. It doesn't necessarily have to be elaborate or consume 20%, 30%, or 50% of their resources, but it needs to say, "Look. Here is an area of the organization where we're going to experiment with these tools in these areas. If you have your own area that you want to experiment, come to us. Here's how you register an idea with us so that we can get the right data and tools in front of you." Not being able to do basic portfolio-like activities around where you're experimenting, I think, is a key area. Then being able to take those and marketing them back to the business
sponsors and the executives saying, "Look. This is where we're experimenting. This is where we're starting from. These are the impacts that we want to see. And here are some of the learning activities that we're doing so that we can see what's happening in our industry or maybe learn from some other industries." I always say CIOs have to be one of the champions of being lifelong learners, and so going out there and seeing what other companies are doing is part of this equation as well. I say something very similar. I say it takes
a village. As a CIO, it takes a village. You can't do it by yourself. Your CIO network is immensely important, especially now with these new innovations that are coming about. But here's the risk that I'll also put on the table if you don't follow Isaac's advice of putting that vision in place—you mentioned that earlier—and really thinking about what's your plan of attack. I'm already seeing this play out in organizations where the edicts are starting to be discussed. What I mean by that is you get the CEO, you get the board, the board, edicts from the board for technology. That's
really bad. We saw this with cloud. We're still dealing with it with cloud today. But the last thing you want is a governance body – whether that's the ELT (executive leadership team) or the board of directors – dictating what technology should be used. The CIO needs to be, absolutely needs to be, in front of that and driving that conversation. If they're not driving that conversation, I think that opens a whole other conversation about where they are and where the organization is today. But I would encourage every CIO to get in front of this and have that plan of attack and communicate it and socialize it with the ELT. But how do we get there? Isaac, how do we get there? CIOs have to get used to putting out something with incomplete information and where the information is going to change. This is actually one of the easier areas because of that.
If Microsoft or Google or somebody else comes out with a big capability that's another game-changer, guess what, board, guess what ELT, the rules of the games have changed again, and we can go modify our strategy. This is actually one of the easier areas for CIOs to put something out there that may end up being wrong. I think the problem here is that CIOs are afraid to be leaders around it, to go and bring all the general managers in place, to bring in the functional leaders in place, and be that coordinator of where the focus should be. I think that's the skillset, that's the job here that's missing. The other side of it, again, is controlled experimentation. It's hard enough to go to a team of 100 Ph.D. data scientists and ask them to have a disciplined, scientific process about how
they're focusing on which problems to focus on, when to pivot, when is it delivering value. It's a constant churn to bring their models from early stages to something that's ready for production. Now you're doing it with the entire organization, like everything is out there. So,
is it worth experimenting in marketing around content creation? Well, if you only put out two blog posts a month, maybe not. Or maybe give a simple tool that's licensed that they can go use. I think the other thing we haven't covered here, Tim, is just about every platform that any enterprise is using is probably embedding a language model capability built into it. I'm not going to name names, but every one of them has it. It's a limited capability because it's often siloed to the data that it has access to. But it's an easy area to experiment, whether it's customer service or CRM or coding or marketing automation. Go tell me what you learned
by experimenting with the technologies that we already have that have AI capabilities built into it. Then go back to your data strategies. Most of our data strategies have focused on structured information. How do we get our rows and columns in an organized way, well-documented, in a catalog so people can do BI and machine learning off of it? We have ignored our unstructured data because it's been hard. Now we have tons of this information. Now the question is, now that there is an LM—and maybe it's a little hard to use it today, but in a year or two, it's going to be easier and easier—where are we going to get strategic value by bringing in content and making it easier for people to ask questions around it? We actually have a question from Arsalan Khan, who always asks these great questions on CXOTalk. Tim, maybe I'll direct this one to you to start. He says, "As IT becomes a commodity—" That's
an interesting point in and of itself that we could argue about. Is AI becoming a commodity? "—and AI becomes used more and more in different areas of IT, will we have an IT department that just has a far diminished role and really is only asked to come into play when there's a functional business need from a user?" Really, he's bringing into question the very role and existence of the IT department. Tim, any thoughts on that one? If that is the mindset of the organization, full stop, I think that's just the beginning of the end for that company. I strongly believe that IT and technology is the ultimate
differentiator between competitive companies, and it gets down to data and how you use data. I don't see the value in IT as a cost center or as just a response to a specific business need. But in fact, if you were to play this out, I actually think the better organizations (broader organizations, not just IT) is where the CEO and the CIO have an incredibly tight relationship. Now, that doesn't happen very often. That's actually fairly rare today. But in those organizations, what happens within IT is pretty magical, and it's very well engrained in the business, the business strategy, and the direction that the company is taking, especially within their industry and starts to impact even economies based on their success. I don't look at it as IT is potentially heading toward this kind of somewhat nonexistence—I know that wasn't the word that was used—toward this diminished capacity. I actually see
it going the other direction, and I think that we already have good examples of that in some large enterprises today. We just need to expand that leadership capability and learn. This is a bigger question than this show about generative AI, but CIOs need to learn how to be better leaders. I'm going to say it. That's one of the biggest challenges I think that we have is that, historically, IT has actually been struggling with this.
I think the rubber is hitting the road now and we've got to step up to the challenge. This is why you're starting to see some CIOs get sidestepped and get diminished in their capability because they're not able to step up to the challenges that companies need them to today. And we see other roles that are starting to come into play to augment that. I do think there is a bigger opportunity for those that are willing to A) understand it, B) learn from it, and C) really have the courage to go after it. We're using the word "CIO" here, but I think the real magic also happens when the CIO has a team of leaders underneath them. You talk about where CIOs are spending their time to really understand the impacts and to market the solutions. They're out there customer-facing. They're out there leadership-facing. They need to change the mindset and culture. That means they need
really good lieutenants who are learning how you connect problem and opportunity to solutions. The people doing this are what I call digital trailblazers, and these are people that, depending on the state of technology, are moving up stack, are defining solutions at a much higher level than we have ever done before. We're not defining storage solutions anymore. We're using cloud capabilities to do that. What AI is going to be doing over the next three years is maybe we're doing a lot less coding and a lot more prompting for code solutions, code examples, to help me document the code, help me build an API out that's a little bit more robust.
We're still required to move up stack, continue to bring problem and opportunity into solutions. There really is one major change as these technologies get easier to use, and that is, we're taking capabilities that we used to have to go to IT for and now they're becoming business capabilities. We're not just doing spreadsheets and Word documents and Office 365. We're now doing analytics. We're doing small AI out there. We're doing code development with low-code platforms. We're taking things that we didn't have enough people in IT to do over long periods of time, and we're bringing it out to our marketers, our finance people, and our operations, and saying, "You have a ton more capability to go do over here."
I don't think IT is going anywhere any time soon. I think the challenge is learning these new capabilities, and it's hard. We're not just coding anymore. We're connecting an idea to dozens if not hundreds of different ways to stitch a solution. And now we're trying to figure out what's a good or optimal way to start experimenting and then seeing which are the solutions that will help us get to market with a technology and a capability a lot faster. We have a question from Twitter. The question is, "What about the data privacy? How do you prevent or mitigate the leak or misuse of intellectual property and confidential information?" This is a risk. This is a concern. It's already happening. We're already seeing examples of this.
I just wrote a blog post about an example of IROs (investor relations officers) that are using ChatGPT with pre-published financial information to build their press releases, their financial releases. That's a concern. Then, of course, you see it in Copilot with code creation. I think the biggest opportunity here—and it's going to sound like I'm kicking the can down the road a little bit, but I don't mean it that way—is education. We cannot put the technology guardrails in place to cover every permutation that we ever
are going to run across without inhibiting the ability to experiment and really explore. This is where technology does not replace the human. We need the humans to use that gray matter between the ears and think about what they're doing. And we need to educate them and help them understand where the risks are so that they can make good, educated, valued decisions from their seat. I think that's incredibly important. I think too often we think of IT people (or others think of IT people) as machines that go to the freezer and get the box. We've way over-rotated on that, and we need to come back to "these are people." These are people that have a brain, that can think, and we need to give
them the trust and build that education so that they can help make great business decisions, whether they're the most senior person in the IT org or the most junior person in the IT org. Those organizations that have gone down that path have had wild success at limiting the risks. Also, these people start to become identifiers to say, "You know what? I recognize this issue.
Maybe we should talk about that," and educate the rest of the team. They become lampposts out in the distance, too. I think that's another piece that we often don't think about is the power of the organization. To me, that's where the guardrails start.
How are attitudes in the C-suite, in general, evolving towards AI, towards the consumption of AI, and the adoption of AI? Where are we today, Isaac? We have a spectrum. I think some believe it's science fiction and we're ready for Star Trek – you know, "Computer, help me with this problem." – and expect business leaders and IT leaders to collaborate to get an answer around that. I think you have a lot of naysayers who would prefer keeping business as usual, prefer using the tools as they exist.
I see that as a spectrum that we see with every new capability. The main difference here is it's moving a lot faster and a lot more people are experimenting with it, so it needs to come together much faster. I think how we consume generative AI will evolve. Isaac talked a lot about building LLMs within your organization. That's
really challenging and really expensive to do. One organization recently mentioned—this is a vendor—that they're paying $0.5 million a year for junior folks. And the more senior folks, they're paying $1 million a year. Now that's an expensive individual, and if you go back to when we were
trying to put data scientists within our IT org, we were having trouble finding that skillset. I think there's that. There's also the actual cost of running these models in the cloud and paying for that. I think it's just too complicated for most organizations to do it themselves. I do think that the way we consume this will be through those enterprise apps that we already are using because they understand the data, they characterize the data, they can map the data and market, and they can help put the right guardrails in place. It's
a much easier lift for the average enterprise. There will still be corner cases where folks are building their own LLMs and managing that. But those will be corner cases and very, very specific. It will not be the masses. I agree with you. It's really expensive.
The skillset to work with LLMs is hard to get. This is really about, number one, getting your data understood so that when we start embedding them in LLMs, you know what that data looks like, and then training and educating the employees to the executives about planning for the art of the future where you can ask those questions and there may be an LLM in the middle that's being able to answer them. It's not really necessarily about going out and building it yourself. The other thing we haven't talked about, Michael, and we don't have time today, but I would encourage folks to just be aware of the whole regulatory space. There is a ton
of regulation that is on the books today as well as in process at a state, federal, and global level that companies have to contend with, and they have to be aware of. It's only going to get messier as we go through time. Again, the problem gets to be harder over time (not easier) and more complicated. I think these are things that (going back to your point about education) we have to learn. We have to spend time and learn and learn fast. A couple of comments from Twitter. Doug Gillette on LinkedIn says that the answer to the question about the relationship between AI and employees, he says, "You don't ban AI but monitoring employees is where many should start." I know that can potentially be also a provocative statement.
Then Arsalan Khan comes back on Twitter, and he says, "Do CIOs and CXOs, in general, actually understand the value, the financial value of their data and where to therefore make investments in data?" Very quickly, any thoughts on this data evaluation and investment question? To the first comment about monitoring employee behavior, I think that's a massive hot button and something that most organizations should absolutely stay away from. There are specific ways that IT orgs are doing that, but they're incredibly transparent about how they're doing it. It comes back to trust in your employees. Then to the second question about the guardrails. I do think that there is more conversation to be had there and more understanding to be had there. We're in the very early innings of this. But the game is going very, very fast.
I would add risk and opportunity sides. On the risk side, we talk about chief privacy officers and chief data officers. We're now at a point where every business leader needs to understand the basics of this. Where are experiments happening? What data is being used? Is it being used in an appropriate way? I think we're taking that knowledge and that responsibility and saying, "We need to have a lot more people with that level of knowledge to make better decisions around them." Then on the flip side, they have to take
a stronger role in saying, "We're going to experiment chasing after value and chasing after experiments here. Here's what the value looks like when we want to have developers do something or have a test out in the field with some IoT technologies." Being able to capture that in a way that you can go and say, "We're heading down a path that's starting to deliver value," because if we end up in a situation and said, "What's this added data source worth to us? What's the ROI around an experiment?" we're not going to be able to calculate it upfront. We're only going to be able to say that the experiments we're doing week-to-week, month-to-month are leading us down toward this value equation that we think will ultimately deliver financial returns. It's important to understand something that Isaac mentioned, and I want to underscore this because it's really important to focus on this. Focus on value not cost – value, not cost. The reason I say that is, quite often, people are focused on, "Oh, well, the cost to do something is X," but they don't necessarily understand what the upside of doing that is. To the question about, "Do we understand the
value of data?" No, we don't. We're learning that as we go. I think that will get exposed as we go through time and more experimentation. Deep Khandelwal says, "How can you bridge the knowledge gap in employees to help them get ready for AI and new technologies?" Isaac? I want to market teams that are having success. I want to be able to bring that
to the forefront and show what they're doing, how they're going about doing it. I'm thinking of the old school town hall picture, but we need to be able to show people examples. They went out there. They went above their job. They went after something that had a
real problem, an opportunity statement around it. It was defined upfront. They went through two or three pivots in their experimentation. And here's something. Maybe it's not ready for production, but here's something that proves that they're heading in the right direction.
I think CIOs, with their CHROs, with their CMOs, need to have a program about bringing these success stories across the organization. We've talked about that for a long time, for decades, and I think we've kind of lost the marketing ability. But I agree with Isaac. We need to bring that back, finding those experiments that are successful and being able to support the teams that have discovered that. But also, let's support the people who failed. They tried, they failed, and they moved on to the next thing. What did they learn from it? Let's not just focus on the successes. Let's also remember the failures or
things that we learned from, not the successes. We have another quick question from LinkedIn. This is from Kash Mehdi who says, "Since data is so critical for organizations worldwide, how does one go about creating a data balance sheet? Are you familiar with any particular frameworks or just guidance?" Data balance sheet. There are a couple of books around that in terms of turning data into a product exercise. I think, for most organizations, trying to keep this simple and starting with something that most tools already have today, measuring data quality, measure data utilization, measure the outputs from your data experiments. Where is it leading toward value? You get into quality. You get into being able to load in more data sets efficiently. You're starting to impact the gears that are going to
start impacting results that come after that. The other things you have to think about, especially as you're working at a global level, there are tools that are just emerging onto the market that are starting to consider this. But you also have to think about regulatory and governance and privacy at a global level because that's getting incredibly more complicated, and these tools are what's going to be required to help you through that.
This is from Sylvia Macia. She says, "What is your opinion as to the state of entities using their own internal transformation manager such as portfolio program or project managers as ambassadors or, as McKinsey calls them, analytics translators to help bridge those knowledge gaps between the business functions and the CIO, CTO, and IT organizations?" There are a lot of people with different titles that have that business-IT alignment and business-to-IT translation. Whatever the titles being used inside, I think the most important skill set there is really a product management skill set. Who is the customer? What value, what problem are we solving for them? How are they working today? What are their current tools around it? Why is this strategically important? Every organization has different skill sets around it or different titles around it. But I think the most important thing is to think about a customer focus first and strategic intent second. The last point from Doug Gillette who just wants to add this, he says that it's a complicated tradeoff between AI adoption and monitoring employees. He says,
"Do you give employees free rein and, therefore, refrain from monitoring their activities on AI to maintain the trust? On the other hand, monitoring provides an opportunity to create policies and governance and address issues that might otherwise go unnoticed." Very quickly, Tim, you made a statement about the monitoring versus adoption. Thoughts on that? It's a tough problem. There's no one answer to that question. Each organization is going to be different. You have to think about the culture. You have to think about what you're trying to do, the sensitivity of the data, sensitivity of the risk that comes with it. There are a lot of factors that go into the answer to that, so I wouldn't say that you live at the extremes of whether you monitor or don't monitor. I think it's a more complicated
conversation to have, unfortunately. Isaac, Kash Mehdi comes back and says, "How do CIOs view chief data officers?" I love this. We're getting provocative. I like that. Not as an adversary, as an adversary? If yes or no, how do you empower them as a CIO? You're getting into the entire organizational structures. Does the CDO report to the CIO?
Are they peers? What are they really chartered with? Are they more data governance focused or do they have both data and analytics responsibilities, even though they just have a data officer category? I think it really comes down to sitting down in rooms (or virtual rooms) and coming up with what the objectives are, agreeing with them, and then drawing that line and saying, "Regardless of the reporting structure, here's where the CIO's organization is going to lead things and take ownership of, and here's where the CDO is going to take ownership of." If you try to make too many generalizations around that, you end up with putting people in boxes, which I don't think actually works. Tim, do you want to have a last crack at this? Then you're going to have the last word. I've seen pretty much the spectrum, as Isaac has. I've seen the extremes of the CDO is a fill-in for the CIO that's incapable of taking on some of these more business-centric pieces. Okay, there I said it. Unfortunately,
that's the reality of the world we live in. And so, now we're starting to see chief AI officer. I think the thing that people forget is this title is getting watered down to the point that it's irrelevant. A chief officer role has a very strategic and important role amongst the executive leadership team – or it should. Unfortunately,
we don't necessarily play by those rules evenly. I think there's some conversation we have to get back to doing that. With that, we are out of time. A huge thank you to Tim Crawford and Isaac Sacolick. They are two of the very best CIO advisors out there. Gentlemen, thank you so much for being here. I'm very grateful to you both. Thanks, Michael. Thank you, Michael. A huge thank you to our audience. You guys are intelligent, bright, and sophisticated,
and you make CXOTalk. You guys are awesome. Now before you go, please subscribe to our newsletter, and subscribe to our YouTube channel. Just go to CXOTalk.com. We have amazing shows coming up, live shows where you can ask your questions, and truly, genuinely, you are part of the conversation. So, check out CXOTalk.com. Thanks so much, everybody. Have a great day, and we'll see you again next week. And I should mention, next week, we have the president of the consumer group at T-Mobile on CXOTalk talking about customer experience. The president of the consumer group at T-Mobile, you know he's pretty high up there, so check it out.
Thanks so much, everybody, and we'll see you again next time. Take care.
2023-11-17 16:21