Kazu Gomi has a giant view of the know-how world from his perch in Silicon Valley. And as president and CEO of NTT Analysis, a division of the massive Japanese telecommunications agency NTT, Gomi can management the R&D funds for a large chunk of the fundamental analysis that’s carried out in Silicon Valley.
And maybe it’s no shock that Gomi is pouring some huge cash into AI for the enterprise to find new alternatives to make the most of the AI explosion. Final week, Gomi unveiled a brand new analysis effort to give attention to the physics of AI and nicely as a chip design for an AI inference chip that may course of 4K video sooner. This comes on the heels of analysis tasks introduced final 12 months that would pave the way in which for higher AI and extra power environment friendly knowledge facilities.
I spoke with Gomi about this effort within the context of different issues huge corporations like Nvidia are doing. Bodily AI has grow to be a giant deal in 2025, with Nvidia main the cost to create artificial knowledge to pretest self-driving vehicles and humanoid robotics to allow them to get to market sooner.
And constructing on a narrative that I first did in my first tech reporting job, Gomi mentioned the corporate is doing analysis on photonic computing as a strategy to make AI computing much more power environment friendly.

A long time in the past, I toured Bell Labs and listened to the ambitions of Alan Huang as he sought to make an optical pc. Gomi’s crew is making an attempt to do one thing related a long time later. If they’ll pull it off, it may make knowledge facilities function on rather a lot much less energy, as mild doesn’t collide with different particles or generate friction the way in which {that electrical} indicators do.
Throughout the occasion final week, I loved speaking to a bit of desk robotic known as Jibo that swiveled and “danced” and informed me my important indicators, like my coronary heart fee, blood oxygen stage, blood strain, and even my ldl cholesterol — all by scanning my pores and skin to see the tiny palpitations and shade change because the blood moved by my cheeks. It additionally held a dialog with me through its AI chat functionality.
NTT has greater than 330,000 workers and $97 billion in annual income. NTT Analysis is a part of NTT, a world know-how and enterprise options supplier with an annual R&D funds of $3.6 billion. About six years in the past it created an R&D division in Silicon Valley.
Right here’s an edited transcript of our interview.

VentureBeat: Do you are feeling like there’s a theme, a prevailing theme this 12 months for what you’re speaking about in comparison with final 12 months?
Kazu Gomi: There’s no secret. We’re extra AI-heavy. AI is entrance and middle. We talked about AI final 12 months as nicely, but it surely’s extra vivid at the moment.
VentureBeat: I needed to listen to your opinion on what I absorbed out of CES, when Jensen Huang gave his keynote speech. He talked rather a lot about artificial knowledge and the way this was going to speed up bodily AI. As a result of you’ll be able to check your self-driving vehicles with artificial knowledge, or check humanoid robots, a lot extra testing will be carried out reliably within the digital area. They get to market a lot sooner. Do you are feeling like this is smart, that artificial knowledge can result in this acceleration?
Gomi: For the robots, sure, 100%. The robots and all of the bodily issues, it makes a ton of sense. AI is influencing so many different issues as nicely. In all probability not every little thing. Artificial knowledge can’t change every little thing. However AI is impacting the way in which firms run themselves. The authorized division is perhaps changed by AI. The HR division is changed by AI. These sorts of issues. In these situations, I’m undecided how artificial knowledge makes a distinction. It’s not making as huge an affect as it might for issues like self-driving vehicles.
VentureBeat: It made me suppose that issues are going to come back so quick, issues like humanoid robots and self-driving vehicles, that we now have to determine whether or not we actually need them, and what we would like them for.
Gomi: That’s a giant query. How do you take care of them? We’ve positively began speaking about it. How do you’re employed with them?

VentureBeat: How do you utilize them to enrich human employees, but in addition–I believe certainly one of your folks talked about elevating the usual of dwelling [for humans, not for robots].
Gomi: Proper. Should you do it proper, completely. There are numerous good methods to work with them. There are definitely dangerous situations which might be attainable as nicely.
VentureBeat: If we noticed this a lot acceleration within the final 12 months or so, and we are able to count on artificial knowledge will speed up it much more, what do you count on to occur two years from now?
Gomi: Not a lot on the artificial knowledge per se, however at the moment, one of many press releases my crew launched is about our new analysis group, known as Physics of AI. I’m wanting ahead to the outcomes coming from this crew, in so many various methods. One of many fascinating ones is that–this humanoid factor comes close to to it. However proper now we don’t know–we take AI as a black field. We don’t know precisely what’s occurring contained in the field. That’s an issue. This crew is wanting contained in the black field.
There are numerous potential advantages, however one of many intuitive ones is that if AI begins saying one thing flawed, one thing biased, clearly you have to make corrections. Proper now we don’t have an excellent, efficient strategy to right it, besides to simply hold saying, “That is flawed, you need to say this as an alternative of that.” There’s analysis saying that knowledge alone gained’t save us.
VentureBeat: Does it really feel such as you’re making an attempt to show a child one thing?
Gomi: Yeah, precisely. The fascinating superb situation–with this Physics of AI, successfully what we are able to do, there’s a mapping of information. Ultimately AI is a pc program. It’s made up of neural connections, billions of neurons related collectively. If there’s bias, it’s coming from a specific connection between neurons. If we are able to discover that, we are able to in the end cut back bias by reducing these connections. That’s the best-case situation. Everyone knows that issues aren’t that simple. However the crew might be able to inform that in the event you lower these neurons, you may be capable of cut back bias 80% of the time, or 60%. I hope that this crew can attain one thing like that. Even 10% continues to be good.
VentureBeat: There was the AI inference chip. Are you making an attempt to outdo Nvidia? It looks as if that might be very laborious to do.

Gomi: With that exact undertaking, no, that’s not what we’re doing. And sure, it’s very laborious to do. Evaluating that chip to Nvidia, it’s apples and oranges. Nvidia’s GPU is extra of a general-purpose AI chip. It may well energy chat bots or autonomous vehicles. You are able to do every kind of AI with it. This one which we launched yesterday is barely good for video and pictures, object detection and so forth. You’re not going to create a chat bot with it.
VentureBeat: Did it appear to be there was a chance to go after? Was one thing not likely working in that space?
Gomi: The quick reply is sure. Once more, this chip is unquestionably personalized for video and picture processing. The hot button is that with out decreasing the decision of the bottom picture, we are able to do inference. Excessive decision, 4K pictures, you should use that for inference. The profit is that–take the case of a surveillance digital camera. Possibly it’s 500 meters away from the thing you need to have a look at. With 4K video you’ll be able to see that object fairly nicely. However with standard know-how, due to processing energy, you must cut back the decision. Possibly you would inform this was a bottle, however you couldn’t learn something on it. Possibly you would zoom in, however you then lose different data from the realm round it. You are able to do extra with that surveillance digital camera utilizing this know-how. Increased decision is the profit.

VentureBeat: This is perhaps unrelated, however I used to be fascinated by Nvidia’s graphics chips, the place they have been utilizing DLSS, utilizing AI to foretell the subsequent pixel you have to draw. That prediction works so nicely that it obtained eight occasions sooner on this technology. The general efficiency is now one thing like–out of 30 frames, AI may precisely predict 29 of them. Are you doing one thing related right here?
Gomi: One thing associated to that–the rationale we’re engaged on this, we had a undertaking that’s the precursor to this know-how. We spent lots of power and assets previously on video codec applied sciences. We offered an early MPEG decoder for professionals, for TV station-grade cameras and issues like that. We had that base know-how. Inside this base know-how, one thing much like what you’re speaking about–there’s a little bit of object recognition occurring within the present MPEG. Between the frames, it predicts that an object is shifting from one body to the subsequent by a lot. That’s a part of the codec know-how. Object recognition makes that occur, these predictions. That algorithm, to some extent, is used on this inference chip.
VentureBeat: One thing else Jensen was saying that was fascinating–we had an structure for computing, retrieval-based computing, the place you go right into a database, fetch a solution, and are available again. Whereas with AI we now have the chance for reason-based computing. AI figures out the reply with out having to look by all this knowledge. It may well say, “I do know what the reply is,” as an alternative of retrieving the reply. It might be a unique type of computing than what we’re used to. Do you suppose that will likely be a giant change?
Gomi: I believe so. Numerous AI analysis is happening. What you mentioned is feasible as a result of AI has “information.” As a result of you may have that information, you don’t should go retrieve knowledge.

VentureBeat: As a result of I do know one thing, I don’t should go to the library and look it up in a e-book.
Gomi: Precisely. I do know that such and such occasion occurred in 1868, as a result of I memorized that. You would look it up in a e-book or a database, but when you understand that, you may have that information. It’s an fascinating a part of AI. Because it turns into extra clever and acquires extra information, it doesn’t have to return to the database every time.
VentureBeat: Do you may have any explicit favourite tasks occurring proper now?
Gomi: A pair. One factor I need to spotlight, maybe, if I may choose one–you’re wanting carefully at Nvidia and people gamers. We’re placing lots of give attention to photonics know-how. We’re fascinated by photonics in a few other ways. If you have a look at AI infrastructure–you understand all of the tales. We’ve created so many GPU clusters. They’re all interconnected. The platform is large. It requires a lot power. We’re working out of electrical energy. We’re overheating the planet. This isn’t good.
We need to tackle this challenge with some completely different methods. One in all them is utilizing photonics know-how. There are a few other ways. First off, the place is the bottleneck within the present AI platform? Throughout the panel at the moment, one of many panelists talked about this. If you have a look at GPUs, on common, 50% of the time a GPU is idle. There’s a lot knowledge transport taking place between processors and reminiscence. The reminiscence and that communication line is a bottleneck. The GPU is ready for the information to be fetched and ready to jot down outcomes to reminiscence. This occurs so many occasions.
One concept is utilizing optics to make these communication traces a lot sooner. That’s one factor. Through the use of optics, making it sooner is one profit. One other profit is that in relation to sooner clock speeds, optics is far more energy-efficient. Third, this entails lots of engineering element, however with optics you’ll be able to go additional. You’ll be able to go this far, and even a few toes away. Rack configuration could be a lot extra versatile and fewer dense. The cooling necessities are eased.
VentureBeat: Proper now you’re extra like knowledge middle to knowledge middle. Right here, are we speaking about processor to reminiscence?

Gomi: Yeah, precisely. That is the evolution. Proper now it’s between knowledge facilities. The following section is between the racks, between the servers. After that’s throughout the server, between the boards. After which throughout the board, between the chips. Finally throughout the chip, between a few completely different processing models within the core, the reminiscence cache. That’s the evolution. Nvidia has additionally launched some packaging that’s alongside the traces of this phased method.
VentureBeat: I began protecting know-how round 1988, out in Dallas. I went to go to Bell Labs. On the time they have been doing photonic computing analysis. They made lots of progress, but it surely’s nonetheless not fairly right here, even now. It’s spanned my entire profession protecting know-how. What’s the problem, or the issue?
Gomi: The situation I simply talked about hasn’t touched the processing unit itself, or the reminiscence itself. Solely the connection between the 2 elements, making that sooner. Clearly the subsequent step is we now have to do one thing with the processing unit and the reminiscence itself.
VentureBeat: Extra like an optical pc?
Gomi: Sure, an actual optical pc. We’re making an attempt to do this. The factor is–it sounds such as you’ve adopted this matter for some time. However right here’s a little bit of the evolution, so to talk. Again within the day, when Bell Labs or whoever tried to create an optical-based pc, it was principally changing the silicon-based pc one to 1, precisely. All of the logic circuits and every little thing would run on optics. That’s laborious, and it continues to be laborious. I don’t suppose we are able to get there. Silicon photonics gained’t tackle the difficulty both.
The fascinating piece is, once more, AI. For AI you don’t want very fancy computations. AI computation, the core of it’s comparatively easy. Every little thing is a factor known as matrix-vector multiplication. Data is available in, there’s a consequence, and it comes out. That’s all you do. However you must do it a billion occasions. That’s why it will get sophisticated and requires lots of power and so forth. Now, the great thing about photonics is that it may well do that matrix-vector multiplication by its nature.
VentureBeat: Does it contain lots of mirrors and redirection?

Gomi: Yeah, mirroring after which interference and all that stuff. To make it occur extra effectively and every little thing–in my researchers’ opinion, silicon photonics might be able to do it, but it surely’s laborious. You need to contain completely different supplies. That’s one thing we’re engaged on. I don’t know in the event you’ve heard of this, but it surely’s lithium niobate. We use lithium niobate as an alternative of silicon. There’s a know-how to make it into a skinny movie. You are able to do these computations and multiplications on the chip. It doesn’t require any digital elements. It’s just about all carried out by analog. It’s tremendous quick, tremendous energy-efficient. To some extent it mimics what’s occurring contained in the human mind.
These {hardware} researchers, their objective–a human mind works with perhaps round 20 watts. ChatGPT requires 30 or 40 megawatts. We will use photonics know-how to have the ability to drastically upend the present AI infrastructure, if we are able to get all the way in which there to an optical pc.
VentureBeat: How are you doing with the digital twin of the human coronary heart?
Gomi: We’ve made fairly good progress during the last 12 months. We created a system known as the autonomous closed-loop intervention system, ACIS. Assume you may have a affected person with coronary heart failure. With this technique utilized–it’s like autonomous driving. Theoretically, with out human intervention, you’ll be able to prescribe the appropriate medication and therapy to this coronary heart and convey it again to a traditional state. It sounds a bit fanciful, however there’s a bio-digital twin behind it. The bio-digital twin can exactly predict the state of the guts and what an injection of a given drug may do to it. It may well rapidly predict trigger and impact, determine on a therapy, and transfer ahead. Simulation-wise, the system works. We have now some good proof that it’ll work.

VentureBeat: Jibo, the robotic within the well being sales space, how shut is that to being correct? I believe it obtained my ldl cholesterol flawed, but it surely obtained every little thing else proper. Ldl cholesterol appears to be a tough one. They have been saying that was a brand new a part of what they have been doing, whereas every little thing else was extra established. If you may get that to excessive accuracy, it might be transformative for the way typically folks should see a physician.
Gomi: I don’t know an excessive amount of about that exact topic. The standard manner of testing that, in fact, they’ve to attract blood and analyze it. I’m certain somebody is engaged on it. It’s a matter of what sort of sensor you’ll be able to create. With non-invasive gadgets we are able to already learn issues like glucose ranges. That’s fascinating know-how. If somebody did it for one thing like ldl cholesterol, we may carry it into Jibo and go from there.