Yotta CEO Sunil Gupta on Supercharging India’s Fast-Growing AI Market – Ep. 225

AI transcript
0:00:00 [MUSIC]
0:00:10 >> Hello, and welcome to the NVIDIA AI podcast.
0:00:13 I’m your host, Noah Kravitz.
0:00:15 We’re recording at GTC24 in San Jose,
0:00:18 California, and we’re here to talk about
0:00:20 the fast growing AI market in India.
0:00:23 My guest is Sunil Gupta.
0:00:25 Sunil is the co-founder,
0:00:26 managing director and CEO of Yata Data Services.
0:00:30 Yata is the first Indian cloud service provider
0:00:32 member of the NVIDIA partner network program,
0:00:35 and the company Shakti Cloud offering
0:00:37 is India’s fastest AI supercomputing infrastructure,
0:00:41 featuring 16 exaflops of AI compute capacity
0:00:44 supported by more than 16,000 NVIDIA H100 GPUs.
0:00:49 Sunil has been a busy man this week,
0:00:51 but he’s taken some time to stop by
0:00:53 the podcast studio here at GTC to tell us all about
0:00:57 Yata and its role in India’s fast growing AI sector.
0:01:01 So let’s get right to it.
0:01:02 Sunil Gupta, thank you so much for stopping by.
0:01:05 Welcome to the NVIDIA AI podcast.
0:01:07 >> Thank you for having me. Thank you.
0:01:08 >> So let’s start with the basics for listeners
0:01:11 who maybe just heard of Yata for the first time this week.
0:01:13 You’ve been in the news, a lot to congratulate you on.
0:01:16 What is Yata? What does the company do?
0:01:18 >> Great. So Yata is the managed data center and
0:01:21 cloud service provider operating in India for the last five years.
0:01:24 We have been running our own self-designed and
0:01:28 engineered and constructed data centers.
0:01:31 We that way offer traditional data center,
0:01:33 co-location, managed hosting cloud,
0:01:35 and a variety of managed services there.
0:01:38 We have been offering GPU services for the last four years.
0:01:42 Ever since we started, but those were the smaller GPUs,
0:01:44 the A40s, V40s, and T4s,
0:01:46 and the use cases used to be creating
0:01:49 content and game creations and things like that.
0:01:52 Possibly those were the credentials that I have so
0:01:55 much of large data center campuses,
0:01:56 and I have got the experience of handling GPUs,
0:02:00 and delivering it to enterprise customers for different use cases,
0:02:03 and from an Indian scale point of view,
0:02:05 if I’m having 700 GPUs today in my data center,
0:02:08 that still possibly is the largest deployment of GPUs in India,
0:02:11 and those were the credentials which
0:02:12 possibly attracted the eye of NVIDIA India,
0:02:14 and they suggested to Jensen that possibly these are the right guys,
0:02:17 having data center that have the power and the right skill sets,
0:02:20 and exposure to GPUs that,
0:02:23 and then they are there as I did in one of my interviews just recently,
0:02:26 that I’m hungry and ambitious,
0:02:28 and I want to take a plunge of in this market,
0:02:32 which possibly can become the largest or one of
0:02:34 the largest market in the world,
0:02:35 and it can also become a garage as
0:02:38 a service provider to the rest of the world for developing AI models.
0:02:41 So, we took the jump, and today, we have got our first set of deliveries already,
0:02:49 and possibly by end of March,
0:02:51 I’ll have all the deliveries completed,
0:02:53 and by around 15th May,
0:02:54 we are targeting to go live giving customers our GPU services.
0:02:58 >> Fantastic. 15th of May, you see?
0:03:00 >> Yeah.
0:03:00 >> Is that Shakti Cloud to go live?
0:03:03 >> That’s Shakti Cloud, right.
0:03:04 >> So, can you tell us about Shakti Cloud?
0:03:05 >> Yes. So, first, yeah.
0:03:07 So, we are running data centers and other cloud and managed services for four years,
0:03:11 we have been giving GPUs for four years,
0:03:13 but yes, Shakti Cloud is something which is essentially a GPU-based cloud.
0:03:18 It has got about 16,384 to be precise,
0:03:22 800 GPUs, it has got a couple of thousands of L40s GPUs,
0:03:28 which are mainly for inferencing purposes.
0:03:30 We have put up this entire GPU Cloud on
0:03:33 NVIDIA’s reference architecture to the T,
0:03:34 we are not debating from that even for a single element of that,
0:03:40 it does not debating from that.
0:03:41 So, essentially, there’s an InfiniBand layer,
0:03:43 which is actually in a leaf and spine architecture.
0:03:45 So, we are creating pods of 2,000 GPUs,
0:03:50 256 nodes put together, create one pod,
0:03:54 and there’s a core layer on the top of that,
0:03:56 which essentially means that I can connect eight such pods of 2,000 GPUs
0:04:01 into one super pod of 16,800 GPUs.
0:04:04 And that essentially means that on one hand,
0:04:07 I can give to a small startup, a single GPU or a single node
0:04:12 or even a partial element of the GPU,
0:04:16 that is one end of the capabilities for some use cases.
0:04:18 But on the other hand, if some last-scale customer comes and he says,
0:04:22 “My model requires 16,800 working parallel to train a very large language model,”
0:04:28 we can develop even that as well.
0:04:29 So, that is essentially one element in terms of the underlying processing capability.
0:04:35 Then, you know, you need as much high-speed special type of storage.
0:04:39 You know, the data need to be trained on the GPU,
0:04:42 so it has to be on a high-speed storage.
0:04:44 So, that is what we have put in from Vekka.
0:04:47 Then, what we have done, we have put a complete NVIDIA AI enterprise stack,
0:04:51 software stack on the top of this.
0:04:53 We have, because we have been working to develop our own sovereign cloud in India
0:04:57 for the last two years, so that experience came in very handy.
0:04:59 We, a lot of that coding, actually, we sort of repurposed
0:05:03 to develop our own orchestration layer, our own self-service portal.
0:05:07 And so, today, my self-service portal has got entire NVIDIA AI software stack,
0:05:14 you know, bundled into that.
0:05:16 I’m putting a whole lot of open-source software libraries.
0:05:20 I’m putting in a capability for people to invoke the models from, let’s say,
0:05:25 a hanging face and then bring it into my orchestration layer,
0:05:28 and then people to bring their own data, you know, annotate the data, clean the data,
0:05:32 maybe create more synthetic data, do all that stuff.
0:05:35 It’s essentially MLops operations, and then use an existing model in my marketplace,
0:05:41 which is there in the portal, to fine-tune their own model, right?
0:05:44 And once they fine-tune their model, then they can put that model, you know,
0:05:48 directly for an influencing purpose in the marketplace,
0:05:50 or they can take it to wherever they want to take it,
0:05:51 or they can build that model through APIs into one of their enterprise applications.
0:05:55 So, essentially, what I have tried to do in Shakti Cloud is that on one side,
0:05:59 you are creating a very, I would say, reasonably large-size underlying
0:06:04 infrastructure layer of compute and storage and network.
0:06:07 And on the top of that, you are trying to create a complete, I would say,
0:06:11 you know, self-contained orchestration layer, where users can come online,
0:06:16 they can consume GPUs from one GPU to a partial GPU, thousands of GPUs.
0:06:20 They can decide to create, you know, their own clusters.
0:06:24 So, I give a capability to them to make online Kubernetes clusters,
0:06:27 or they can go for SLUM as a cluster.
0:06:29 And once they have done that, and then, essentially, they bring in the data
0:06:33 and train the models, either they make a model grounds up,
0:06:37 or they use one of the pre-existing models in the marketplace,
0:06:39 and then fine-tune that with their own data.
0:06:41 So, that is essentially what we are trying to do in Shakti Cloud.
0:06:45 Can you tell us a little bit, and for the audience, for myself,
0:06:48 what the digital economy, what the AI scene is like in India,
0:06:53 and what, you know, you talked about building a sovereign cloud,
0:06:56 and clearly, you know, Shakti Cloud is offering, as you said,
0:07:00 services ranging from a partial GPU all the way up.
0:07:03 What does it mean to the Indian economy, to the tech sector in India?
0:07:09 You know, what is this presence going to do?
0:07:11 Well, see, as you know, I just delivered a speech in the conference today,
0:07:15 and the topic of my speech was that it is AI from India, for India,
0:07:21 and also for the world, right?
0:07:22 So, essentially, there were two elements to my speech.
0:07:24 One was that India itself is potentially going to be the world’s first
0:07:29 or the second or third largest market,
0:07:31 and there are some reasons why I am saying so.
0:07:33 And second part of this is was that, just like India has dominated the world scene
0:07:38 as the garage for delivering IT software and services for the last three decades,
0:07:42 there is no reason to believe that India cannot be also an AI garage for the world, right?
0:07:48 What India has already, if I just give you some dynamics on the first part,
0:07:52 why India has the potential to become one of the largest market in the world,
0:07:55 see, you just see the scale of digital adoption.
0:07:57 Today, India has got 900 million internet users,
0:08:01 possibly more than the population of some continents, right?
0:08:04 Out of this 900 million, 600 million are smartphone users,
0:08:07 essentially the users who are consuming videos,
0:08:10 who are consuming images, who are generating videos,
0:08:12 generating images and actually pushing it out, right?
0:08:15 India today is far ahead in terms of adopting digital payments.
0:08:19 You are doing 100 billion payment transactions per month,
0:08:23 which is like 10 times bigger than any other economies in the world.
0:08:26 Right.
0:08:27 India has got more than 5 million trained IT professionals.
0:08:31 This count keeps on increasing and out of which I think 420,000 plus
0:08:34 are actually AI trained professionals.
0:08:35 Okay.
0:08:36 This is the skill set side and the economic side.
0:08:38 By the way, Indian economy is growing,
0:08:40 which is the bright spot in the world scene.
0:08:42 It’s 7.5% CAGR and as per all the various reports for the next few years,
0:08:47 because the demographic dividend which India is enjoying,
0:08:49 the population which is earning is more than the population which is dependent,
0:08:53 India is expected to grow with the same speed for next couple of years.
0:08:57 So when you combine all these factors together
0:08:59 and also couple that with the factor that on the infrastructure side,
0:09:02 where typically India was seen as maybe a laggard economy,
0:09:05 in the last couple of years,
0:09:06 India is building infrastructure like anything in terms of digital infrastructure,
0:09:09 not only the network and the fiber going to the last mile,
0:09:12 in the 5G going to the last mile,
0:09:13 Indian data center scene which was just about 200 megawatt just by 2015,
0:09:19 today India is boasting of around 1 gig of data center capacity,
0:09:23 which is ready and into production and there’s another about 700 megawatt,
0:09:27 which is going to be going live in the next couple of months.
0:09:30 And this is something which is growing at a 40% CAGR.
0:09:33 So and because all these data centers came up just in the last 7-8 years
0:09:37 because of the hyperscalers coming and putting up the shops in India,
0:09:40 so you can just imagine that these data centers have been built
0:09:43 as per the specification of the hyperscalers.
0:09:45 So these are world-class, latest generation data centers as good as anywhere in the world.
0:09:50 And you know, with some retrofitting and extra engineering,
0:09:55 many of these data centers can be customized to handle the GPU load.
0:10:01 So if you see, combine the skill stats, growing economy, the digital adoption
0:10:06 and also that India also is now building up on its infrastructure
0:10:11 and I now couple that with Shakti Cloud getting launched,
0:10:15 India will also have a very large supercomputer for the purpose
0:10:20 of training AI models and putting them for inferencing.
0:10:23 There’s no reason why India will not, first of all,
0:10:26 become a very, very large AI market itself.
0:10:29 We can become a consumer of AI,
0:10:30 but it will also become a very, very big service provider for the rest of the words.
0:10:35 You know, to consume AI from India.
0:10:37 One of the things I noticed in reading about Shakti Cloud
0:10:39 and listening to some of the media coverage
0:10:41 was the importance of self-service.
0:10:44 Can you talk a little bit about that?
0:10:46 About being able to offer a complete end-to-end high performance and AI tune platform,
0:10:51 especially considering the diversity of customers that you’re going to be serving,
0:10:55 you know, startups, academics, researchers, and obviously industry at scale.
0:10:59 Yes. So this is one of the things which we focused on big time.
0:11:03 And I can see, you know, possibly we are following Nvidia’s footsteps.
0:11:06 So if you see Jensen’s keynote also two days back, you know,
0:11:09 while he was announcing the launch of GB200,
0:11:12 which was a hardware part, I can say,
0:11:13 but 70% of his talk actually was focusing on software, right?
0:11:17 And as much capabilities you are giving to the end users, you know,
0:11:21 which is making their life easy to develop their AI model
0:11:24 and put them for inferencing for different industrial use cases.
0:11:27 And as much you make it easy, you know, for them to do this, you know,
0:11:31 and that is where the software comes in.
0:11:33 So as I said that we have been working to develop our own sovereign cloud in India
0:11:38 for the last two years. This is something which is the need of the art for India.
0:11:42 We actually use that capability to also very, very quickly,
0:11:46 possibly just in about three months, we actually developed our own self service portal,
0:11:50 a orchestration layer, completely homegrown,
0:11:52 which essentially means that one end users can come in.
0:11:56 It can be a small startup or it can be a very, very large organization
0:11:59 wanting thousands of GPUs for everybody.
0:12:02 You can just come online, create your own account, authenticate yourself.
0:12:06 If you end user, you can authenticate using my AD.
0:12:09 If you are a large enterprise, we can plug in your AD
0:12:11 and you can authenticate with that.
0:12:13 And once you have done that, you starting from, for example,
0:12:17 you want to have a bare metal node.
0:12:20 So whether you want to have a single bare metal node
0:12:22 or whether you want to have thousands of nodes, you can subscribe it online.
0:12:25 You know, it will show your drop down.
0:12:26 You’ll show the types of GPUs available and you can just subscribe it.
0:12:29 You want to put operating system on this.
0:12:31 So we have got images of very operating systems on that.
0:12:33 Now you want to create a Kubernetes cluster on that.
0:12:35 We want to create a slump cluster on that.
0:12:37 We are giving those capabilities.
0:12:38 You select your master nodes, how much work you want.
0:12:40 And then, you know, the cluster gets made right then and there.
0:12:44 And then you can monitor the various health parameters of your clusters.
0:12:46 The other part is serverless, for example, that especially during inferencing,
0:12:51 how I’m seeing the requirement coming in from our customers
0:12:54 that during training time, possibly they need, you know,
0:12:56 maybe a couple of hundreds, a couple of thousands of GPUs dedicated for themselves
0:12:59 for train their model.
0:13:01 But once they have done that over a, let’s say, a period of 60 days or 90 days,
0:13:04 or whatever the time it takes, depending on the size of their model.
0:13:07 After that, they’re putting the model for inferencing.
0:13:09 And inferencing traffic is like, like any other application put on the internet
0:13:13 where users are coming and consuming your model of the application.
0:13:17 Now that traffic is going to be very, very busted.
0:13:19 There will be peaks where there’s a huge user traffic
0:13:21 and they’ll be trough where basically the traffic is not there.
0:13:24 So essentially, the users are saying that instead of we paying
0:13:27 for a dedicated committed capacity of GPUs,
0:13:29 you rather give it to your GPU on demand,
0:13:31 essentially what we call as a serverless function, you know,
0:13:33 where, you know, multiple users are actually vying for the same GPU capacity
0:13:37 and everybody gets satisfied because there’s enough capacity
0:13:40 and not everybody is peaking at the same time.
0:13:42 So that’s for serverless functionality.
0:13:44 Also, we have put in on the fly, you can develop containers
0:13:46 and put the same GPU capacity for inferencing function as well.
0:13:51 Then what we did is that, you know, while there’ll be certain people
0:13:55 who would like to make grounds up large language models,
0:13:58 they don’t need any of my software other than access to GPUs and clustering.
0:14:01 But majority of the people possibly will be coming
0:14:04 to consume one of the foundational model which is available in the marketplace.
0:14:07 So we have a big marketplace in the auction layer
0:14:10 and they would like to bring in their own data, possibly to clean the data,
0:14:13 annotate the data, maybe create some more synthetic data.
0:14:15 And then they will put the data in one of the pre-existing models
0:14:19 and actually create their own fine tune models.
0:14:20 So that I think will be a much, much bigger market going forward.
0:14:24 Most of the enterprises may not be creating their own foundational models.
0:14:28 They would rather be using some of the foundational models
0:14:30 but putting in their own company-specific or industry-specific data
0:14:34 and create their own, you know, use case-specific models.
0:14:37 So that is the whole environment we have created in this.
0:14:40 So there are a whole lot of pre-trained models available,
0:14:44 whether it is something which is a part of the NVIDIA AI stack,
0:14:47 which I am integrating into my this,
0:14:49 or whether you invoke some of the pre-trained models
0:14:51 from Hugging Face or any of the public platforms.
0:14:53 In fact, many of my startup customers who are creating
0:14:57 their own foundational models by default,
0:14:59 they also are going to put those models also in my marketplace.
0:15:03 So that their end customers can come to a marketplace
0:15:05 and start consuming their models to build their own fine-tuned models.
0:15:08 So if you see this entire capability of this end-to-end software,
0:15:14 there’s no end to that.
0:15:15 I mean, I myself don’t know what all we’ll end up doing in the next one year
0:15:18 because every day is a new discovery as to what all is possible.
0:15:21 What essentially we are trying to do is that, yes, right now,
0:15:25 majority of the GPU usage seems to be the domain of the tech companies,
0:15:31 the startups who are actually trying to train the models,
0:15:34 either a foundational horizontal model,
0:15:36 a language model, or an industry-specific LLM.
0:15:39 But gradually, I will see if the market will start moving more and more
0:15:43 towards fine-tuned models,
0:15:44 especially to the needs of a company or enterprise or an industry.
0:15:47 And that will be put for interesting.
0:15:49 So how we are able to meet the demand of this wider audience?
0:15:53 You know, the enterprises in the wider startups,
0:15:55 instead of just focusing on some of the bulk customer upfront,
0:15:58 and that is where the important this software comes in,
0:16:00 multiple people coming in without any manual intervention
0:16:03 or me trying to do something for them,
0:16:04 which delays the whole thing.
0:16:06 They just come and start consuming whatever they want.
0:16:08 Yeah, yeah.
0:16:09 Does your approach to security and reliability change going from,
0:16:15 you know, data centers sort of of the past, if you will,
0:16:18 to rolling out this large, powerful GPU cloud
0:16:23 that’s serving so many diverse, you know, types of users and use cases.
0:16:27 Are there specific things for kind of the AI age
0:16:31 that you have to take into account when it comes to security?
0:16:34 Mostly not.
0:16:35 I mean, and again, I can keep on saying
0:16:37 that we are all ourselves into a learning phase
0:16:39 and we are everyday discovering what are the new dynamics coming in.
0:16:41 But as far as my understanding as on today’s concern,
0:16:44 there’s not a major difference.
0:16:45 Because one is that, yes, you know,
0:16:48 that when you are training the model
0:16:50 or even when you’re putting the model for fine tuning or for inferencing,
0:16:53 you know, there was a specific concern of the customers
0:16:56 that when I’m putting my data, you know,
0:16:58 it is going to be sometime a very, very secretive,
0:17:01 you know, very, very sensitive data or sensitive data for a company,
0:17:04 they want to be doubly sure that even the operator that is ourselves
0:17:08 are not able to have access to the data.
0:17:10 So that is where NVIDIA in their HGX H100 boxes
0:17:13 have come out with a confidential VM feature.
0:17:16 That is something which gives a confidence to the end users
0:17:18 that, you know, my data is not exposed.
0:17:19 And that was very, very important, you know.
0:17:21 This is one key thing which is specific to AI,
0:17:23 specific to GPUs, which NVIDIA also brought in
0:17:25 and which is very, very helpful.
0:17:27 And many of my customers actually asked for this feature.
0:17:29 But if you go beyond this, once you have trained the model
0:17:33 and you put the model for inferencing,
0:17:35 either the model directly goes to inferencing
0:17:37 with some console layer which users can consume the model into
0:17:40 or you plug it into one of the enterprise application.
0:17:42 I think after that it becomes like any other enterprise application
0:17:45 which end users are coming and consuming.
0:17:47 So you will have good users who are genuine users
0:17:49 and you have bad users who are trying to attack it
0:17:51 and who are trying to make it, you know, go bad.
0:17:54 So from that point of view, all the security services
0:17:57 and I’ve got a big practice of cybersecurity,
0:17:59 you know, as a service for my end customers
0:18:01 who are today actually hosting there maybe SAP ERP
0:18:04 or Oracle or some of the intranets and portals.
0:18:07 So for that, you know, many of my existing services
0:18:10 is like I give, you know, a SOC as a service.
0:18:13 There’s a SIEM and SO2.
0:18:14 So there are firewalls as a service.
0:18:15 There’s an IDS/IPS as a service.
0:18:17 You know, users want their end users to be coming
0:18:20 and they should be authenticated.
0:18:21 So there’ll be a sort of a PIM layer or a PAM layer.
0:18:23 So there are a whole lot of security layers.
0:18:25 There’s a DDoS layer, for example.
0:18:27 So yes, for while I’ve sized up my internet backbone
0:18:31 which connects my data center to the rest of the world through internet
0:18:34 for a particular use cases which were present till date.
0:18:36 But now with AI coming in
0:18:38 and once some good, serious good models are put to inferencing use,
0:18:43 I presume the traffic which will be coming
0:18:45 and into my data center to use this model will be humongous.
0:18:48 So from that point of view, expanding and then increasing
0:18:52 the size of my internet backing make it much more robust
0:18:55 and also then also protecting it with a layer of follow-alls
0:18:58 and the DDoS protection layer is something which we are enhancing.
0:19:01 But if you ask me fundamentally, you know,
0:19:04 that the approach to security remains the same
0:19:06 as you would put for an enterprise application.
0:19:07 A certain specific thing which is specific to AI,
0:19:09 like I talked about confidential VM feature in H100,
0:19:12 is something which is new.
0:19:13 Right.
0:19:14 I’m speaking with Sunil Gupta.
0:19:15 Sunil is the co-founder, managing director
0:19:18 and CEO of Yata Data Services.
0:19:20 Yata Data Services is the first Indian cloud service provider member
0:19:24 of the NVIDIA partner network program.
0:19:27 And we’ve been talking about Yata Shakti Cloud
0:19:30 which is launching, well, by the time you hear this,
0:19:32 it may be launched already,
0:19:33 offering India’s fastest AI super computing infrastructure
0:19:37 on end-to-end environment for basically everything you want to do,
0:19:41 whether you’re a startup, huge player already in India,
0:19:43 and as you said, Sunil, beyond providing AI services out to the world.
0:19:48 You’ve been in working with data centers for a while.
0:19:51 As AI has started to explode into the mainstream consciousness
0:19:55 and sense generative AI in particular
0:19:57 has really captured people’s imagination,
0:19:59 there’s been more of a light shine on the sustainability of data centers.
0:20:04 The power, you know, as compute grows,
0:20:07 power needs grow and these data centers grow physically.
0:20:09 How does that sort of coexist with a world
0:20:14 where we’re talking about climate change
0:20:16 and we’re talking about sustainability
0:20:18 and where is the power coming from
0:20:19 and clean power and unclean power and all that kinds of things.
0:20:22 What are your thoughts?
0:20:23 What’s your approach to building a giant data center
0:20:26 and thinking about how sustainability factors in?
0:20:29 No, no, absolutely.
0:20:30 I think it is a very, very fine balance
0:20:33 between your need for growth as well as the need for sustainability
0:20:37 and how you actually balance it is the real key.
0:20:39 I often say that sometime when you are starting late in the industry,
0:20:43 it is like a boom sometime.
0:20:45 There can be disadvantages to that,
0:20:46 but there are like advantages to that.
0:20:48 So you are able to go to the latest technologies,
0:20:52 right in one shot, you are able to take that jump up front.
0:20:54 And similarly, if they are less certain,
0:20:56 concern is about industry like for data center industry,
0:20:59 data centers are known as power guzzlers.
0:21:00 You know, even today, I think 3% of the world energy
0:21:02 is used by data center project to cross two digits.
0:21:05 If the growth of cloud and AI just keeps on happening,
0:21:09 the way it’s happening.
0:21:10 So in India, because we started the data industry late,
0:21:13 we started scaling it up late and now AI also has come up.
0:21:15 So by default, we are baking in the technologies
0:21:18 or the frameworks where not only we build larger data centers
0:21:23 to handle this type of workload,
0:21:24 but we also take care of the concerns for
0:21:26 that you use lesser amount of power
0:21:28 and also you use the right type of power,
0:21:30 which is the green power.
0:21:31 So those type of things are getting baked into our design
0:21:33 right from day one to just to give you an idea.
0:21:36 So number one, when I designed my data centers,
0:21:38 I designed it at a PUE level, which was less than 1.5.
0:21:41 Now for an Indian tropical environment,
0:21:44 you know, having a PUE of 1.5,
0:21:45 you know, compared to what India used to have traditionally
0:21:48 as 1.8 is something which was good.
0:21:50 It was saving you lots of power.
0:21:51 And now the attempt is by way of adopting
0:21:55 latest technologies like a direct liquid cooling
0:21:57 or immersion cooling,
0:21:57 you can potentially bring this PUE down to 1.2
0:22:00 or possibly 1 also.
0:22:01 So that would be a big,
0:22:03 big, big contribution to environment
0:22:04 when you start using lesser power itself, right?
0:22:07 And that’s what we are trying to do even for Shakti Cloud
0:22:09 that we have right now because the power per rack
0:22:14 from a traditional 6 kilowatt per rack,
0:22:16 now you are handling a 48 to 60 kilowatt per rack,
0:22:19 there’s so much of power being put into the same rack.
0:22:22 So we have actually taken the liquid,
0:22:24 the water right up to behind the rack,
0:22:25 which is called a rear door heat exchanger,
0:22:28 you know, which reduces the PUE
0:22:30 and also makes sure that I’m able to handle that workload.
0:22:32 That’s what I’ve done for the Shakti Cloud
0:22:33 environment in one of the floors in my data center.
0:22:36 In my second phase of deployment,
0:22:38 I’m already getting the direct liquid cooling,
0:22:41 which is potentially bringing the PUE to 1.2.
0:22:44 And once NVIDIA certifies,
0:22:45 and that is something which I hope NVIDIA will,
0:22:48 I potentially would like to use immersion cooling
0:22:50 because immersion cooling is something
0:22:52 where you are just dipping the chips
0:22:54 directly into a liquid or water.
0:22:56 And there are no fans and practically the PUE becomes one,
0:23:02 which essentially means that you are not spending
0:23:04 any extra power for cooling the equipment, right?
0:23:07 So this is something which is in terms of you
0:23:10 trying to reduce as low a power as possible.
0:23:12 The second part is the type of power you are using,
0:23:15 are you using more of a coal or thermal based power
0:23:17 or you are using green power?
0:23:18 So the good part is that even today in my data centers,
0:23:22 I’m using more than 50% of the power
0:23:24 which I’m using in my data center,
0:23:25 actually are green sources, it’s lots of hydro sources,
0:23:27 a lot of solar and wind sources.
0:23:29 And because our approach since starting when we started,
0:23:32 Yorta was to build last scale data center campuses,
0:23:35 which can have multiple buildings,
0:23:37 and I can serve customers with megawatts
0:23:39 and megawatts of power and racks.
0:23:41 So we took some fundamental steps.
0:23:43 We actually took power distribution licenses in India.
0:23:45 If you want to have your own power distribution,
0:23:47 you have to take some licenses in the government.
0:23:49 So I ended up taking those distribution licenses
0:23:51 for both my Mumbai and Delhi campuses.
0:23:53 Now that is giving me a leverage to decide my sources of power.
0:23:57 At any more time, I can decide which type of power I consume.
0:24:00 So we actually decided to have more and more element of green
0:24:04 into our power consumption.
0:24:06 And yeah, so I would say that as we have more and more GPUs
0:24:11 into our data center,
0:24:12 the amount of power we’ll be consuming in the buildings
0:24:15 is just going to scale up much faster than I had imagined earlier.
0:24:20 So my race to put my source of power from a 50% to hopefully 100% green
0:24:27 is something going to become faster.
0:24:29 So possibly in the very short period of time,
0:24:31 I would ideally love to have my power source to be completely 100% green.
0:24:35 Great.
0:24:35 You’ve spoken about it a little bit in this conversation already,
0:24:38 but kind of just to put a point on it,
0:24:40 what does Yorta’s partnership with Nvidia mean,
0:24:43 kind of both to the company,
0:24:44 but also to Yorta’s position sort of now
0:24:47 and going forward in the global tech landscape?
0:24:50 Well, today, if you’re talking AI, you’re talking Nvidia, right?
0:24:53 Nvidia is practically holding I think 88% of the whole market share,
0:24:58 right?
0:24:58 And with the type of announcement
0:25:01 which Jensen is making both on the hardware front,
0:25:03 the GB200 is just taking it so many notches
0:25:06 about against competition.
0:25:07 And of course, the bigger focus is clearly that
0:25:11 how do you make the GPUs to the practical use cases of the industry
0:25:14 and you guys are doing so much in terms of the software libraries
0:25:19 which you are bringing in.
0:25:20 So that capability that I’m aligned with Nvidia
0:25:24 and you are becoming a NCP Nvidia,
0:25:27 you know, the cloud partner
0:25:29 and you’re following the reference architecture to the T.
0:25:32 That is something which is giving a huge confidence to the end customer
0:25:35 that these guys are bringing in exactly what Nvidia is doing
0:25:38 in their own VGX cloud.
0:25:39 I’m practically replica of that.
0:25:41 In fact, it is Nvidia’s professional services team.
0:25:44 You know, so first I engage the Nvidia design team
0:25:47 to design my entire GPU cloud.
0:25:49 Now it’s the Nvidia professional services team
0:25:51 who will be landing in India in the first week of April
0:25:52 and till 15th of May, they’ll be there
0:25:55 who will actually will be doing the commissioning also
0:25:57 and then they’ll be putting all the CUDA layer
0:25:58 and all the software layer on the top of that
0:25:59 and then they’ll be taking this whole cloud
0:26:01 through a huge, you know, I would say sequence of testing it
0:26:06 and then we’ll be publishing the performance benchmarks on MLPuff.
0:26:10 And we are expecting because we have followed
0:26:12 the exact, you know, reference design of Nvidia
0:26:16 with the same software layer as well.
0:26:17 So my performance benchmarks of Shakti Cloud
0:26:19 ideally should be almost same
0:26:20 as what Nvidia has published for DJ cloud also.
0:26:22 So this is something which in all my conversations
0:26:25 with end customers, the moment I tell them this
0:26:28 coupled with my software layer
0:26:29 and coupled with my managed services
0:26:30 which I’m doing through the end customer in my data center.
0:26:32 This something is going to be huge confidence
0:26:34 that yes, these guys are bringing in absolute top notch
0:26:37 something which is there available in any part of the world.
0:26:41 So that is helping it very, very big time.
0:26:43 I feel this is from my side
0:26:45 but I think if you see the reverse way also for Nvidia
0:26:47 it’s like a marriage of two consenting partners.
0:26:51 Right.
0:26:51 For Nvidia also, I think, you know, the Indian market
0:26:54 is a huge market, you know, not to be ignored.
0:26:58 After US, India has the highest potential
0:27:01 to become the market for AI.
0:27:03 Nvidia also needed a partner who knows the local situations
0:27:07 and who has the right set of capabilities
0:27:09 and the infrastructure
0:27:10 to take the Nvidia capabilities to the end customers.
0:27:12 So I think it is a great situation to be in.
0:27:15 We are talking to a whole lot of customers across segments
0:27:19 whether it is government customers or startups
0:27:21 or enterprises or educational institutions
0:27:22 like IIT is in India
0:27:24 or for that matter a whole lot of customers
0:27:26 from other parts of the world
0:27:28 because scarcity is everywhere.
0:27:29 So whether it is Europe or Middle East
0:27:31 or some of the APEC countries, you know,
0:27:33 I’m getting requirements from everywhere.
0:27:35 And yes, because it is Nvidia,
0:27:37 because it is Nvidia reference architecture
0:27:39 and because I’ve got all the software stack out there
0:27:41 and Nvidia, the company is fully behind your tongue
0:27:44 in our success.
0:27:45 That is something which is coming in very, very handy.
0:27:48 That’s tremendous.
0:27:49 Kind of to wrap up
0:27:50 and again everything you’ve said so far sort of leads into this
0:27:53 but to ask the question, what’s the vision?
0:27:56 What’s the vision for Yota in India globally
0:27:59 once Shakti Cloud is deployed?
0:28:02 What comes next?
0:28:02 Where do you see this going?
0:28:04 And if the answer is like you said,
0:28:06 I don’t know because everything’s changing so fast,
0:28:08 that’s okay too.
0:28:08 Yeah, somehow I believe that you can never be a fixed strategy.
0:28:12 You know, the days of having a fixed strategy
0:28:15 and just keep on working on that strategy are gone.
0:28:16 I think if you’re changing your strategy
0:28:18 even after every three months,
0:28:20 essentially it is not change of strategy.
0:28:22 It’s like responding to the market needs, right?
0:28:24 So you have to change yourself constantly,
0:28:26 unlearn old things and relearning new things
0:28:28 which are relevant to the market.
0:28:30 So if you ask me my thought process today,
0:28:33 is that first is India possibly needs 100,000
0:28:38 or maybe multiple 100,000s of GPUs.
0:28:40 Today this 4,000 to 16,000 look big compared to India scale
0:28:44 but if I see it compared to a US market, it is too small.
0:28:47 But all the dynamics which I talked to you
0:28:50 about the Indian digital adoption,
0:28:52 there’s no reason to believe that India
0:28:54 cannot become a much, much larger market.
0:28:56 Of course.
0:28:56 So how I keep on building the GPU capacity,
0:29:00 the latest generation GPU capacity
0:29:02 of Janssen themselves has committed
0:29:03 that we will be the early cloud service provider
0:29:07 who will get access to GB200.
0:29:09 So I definitely like to have the latest generation of GPUs,
0:29:12 the latest generation of network products
0:29:14 available in Shakti Cloud.
0:29:16 We’ll keep on developing on the softer capabilities.
0:29:19 We’ll keep on having more and more pretend models into this
0:29:22 and there’ll be a whole lot of elements will be coming in.
0:29:24 To our startup customer and also other customers,
0:29:27 we are actually signing contracts with them
0:29:29 that whatever models you are creating,
0:29:31 we are again asking them, requesting them
0:29:33 to publish those models again in my marketplace
0:29:35 so that becomes the overall one ecosystem.
0:29:37 Second part is I’m not going to restrict myself to India.
0:29:40 Of course, after Mumbai, which is my starting point,
0:29:42 I’ll actually put my Shakti Cloud node in Delhi,
0:29:45 where my second campus is.
0:29:46 And as per the needs and demand of the customer,
0:29:49 if at all there’s a location specific needs,
0:29:50 tomorrow nothing stops me to launch it,
0:29:52 let’s say in Bangalore or some other part of the country.
0:29:55 What I am very much interested in
0:29:57 and that is the talks we are having even for our other cloud,
0:29:59 which is Yantra Cloud, my normal hyperscale cloud.
0:30:02 And now I’m not applicable to Shakti Cloud is that
0:30:04 there are customers maybe from data residency concerns point of view
0:30:09 who would like to have the capabilities of Shakti Cloud,
0:30:12 but they would like the GPUs to be available in their country.
0:30:15 So, I in any case in my growth chain as a DC operator,
0:30:19 I already had the vision of constructing data centers
0:30:22 in some of the APEC countries,
0:30:23 some of the Middle East and African countries.
0:30:25 Maybe that will go at a little back burner
0:30:29 because constructing a data center into foreign territory
0:30:32 takes its own sweet time, but for Shakti Cloud,
0:30:35 I can still outflows the underlying data center co-location capacity
0:30:40 from one of the partners in those respective territories.
0:30:42 But then I can definitely have GPUs and this whole
0:30:45 Shakti Cloud layer delivered and implemented
0:30:47 and commissioned still managed from India in the respective countries.
0:30:52 So, essentially, the idea is one, vertically keep on adding up to
0:30:57 the capabilities in terms of software.
0:30:59 Second is expanding more and more and more GPUs
0:31:03 into one data center, two data center and multiple data center in India.
0:31:07 And three, taking the Shakti Cloud as a product
0:31:11 to multiple territories across the world wherever there’s a demand.
0:31:13 So much, there’s so much happening.
0:31:15 So much, so much happening in the future.
0:31:17 It’s an amazing time.
0:31:19 Sunil, for listeners who are digesting this and thinking,
0:31:23 I need to learn more about Yata.
0:31:25 This is something I need to keep an eye on.
0:31:27 Website, there’s been media coverage.
0:31:29 Where should they look online to find out more?
0:31:32 So, you can find more about Shakti Cloud on our website.
0:31:35 The URL is www.shakticloud.ai.
0:31:39 You can read quite a lot of information about this is a website
0:31:43 which has just gone like just three days back in the latest version.
0:31:46 And while right now our online portal where you can access
0:31:51 all the services online is still being potent.
0:31:55 So it may take maybe a couple of days more
0:31:56 before users can start coming in and start consuming the services online also.
0:32:00 But there are enough resources for you to know about Shakti Cloud on ShaktiCloud.ai.
0:32:04 Fantastic.
0:32:05 Sunil Gupta, thank you so much for taking the time.
0:32:08 Obviously, busy, busy times for your congratulations on all of it.
0:32:11 But thanks for taking the little time to come out and talk to the podcast audience.
0:32:14 So they can stay abreast, keep an eye on the burgeoning.
0:32:18 I mean, to everything you said, the Indian market,
0:32:21 the Indian scene set to explode.
0:32:24 And it’s a global world now.
0:32:25 So the benefits will be felt all over.
0:32:27 Thank you so much.
0:32:28 Thank you for having me.
0:32:29 [Music]
0:32:39 [Music]
0:32:49 [Music]
0:32:59 [Music]
0:33:07 [Music]
0:33:17 you

India’s AI market is expected to be massive. Yotta Data Services is setting its sights on supercharging it. In this episode of NVIDIA’s AI Podcast, Sunil Gupta, cofounder, managing director and CEO of Yotta Data Services, speaks with host Noah Kravitz about the company’s Shakti Cloud offering, which provides scalable GPU services for enterprises of all sizes. Yotta is the first Indian cloud services provider in the NVIDIA Partner Network, and its Shakti Cloud is India’s fastest AI supercomputing infrastructure, with 16 exaflops of compute capacity supported by over 16,000 NVIDIA H100 Tensor Core GPUs. Tune in to hear Gupta’s insights on India’s potential as a major AI market and how to balance data center growth with sustainability and energy efficiency.

Leave a Comment