Intel Corp.. is acquiring AI optimization software vendor SigOpt, a move the chip maker said would complement its existing AI software portfolio while integrating SigOptâs tools with its AI hardware to accelerate and scale AI software used by model developers.
The acquisition also addresses the growing complexity of machine learning and neural network models and the resulting inability of hardware to keep pace.
Terms of the transaction announced Thursday (Oct. 29) were not disclosed. Intel (NASDAQ: INTC) said it expects the acquisition to close by the end of this quarter.
San Francisco-based SigOptâs co-founders and brain trust, CEO Scott Clark and CTO Patrick Hayes, will join Intelâs machine learning team.
SigOpt was founded in 2014 to create a commercial product from Clarkâs academic research at Cornell University on Bayesian optimization techniques. Combined with Intelâs AI computing and machine learning capabilities, Clark said SigOptâs optimization software would help âunlock entirely new AI capabilities for modelers.â
SigOptâs AI software is designed to boost productivity and performance across hardwareÂ and software parameters, resulting in more accurate and better performing machine learning modelsâeven as complexity grows.
âSigOptâs AI software platform and data science talent will augment Intel software, architecture, product offerings and teams,â said Raja Koduri, Intelâs senior vice president and general manager of architecture, graphics and software.
The startup previously attracted the attention of In-Q-Tel, the investment arm of U.S. intelligence agencies, which eventually acquired a stake in the AI software developer.
Among the companyâs strengths is its focus on metrics used to improve the performance of machine learning models.
The SigOpt deal therefore addresses concerns raised last year by Naveen Rao, vice president and general manager of Intelâs AI Products Group. Neural networks have grown so big, Rao noted, with so many parameters to calculate, that AI hardware is unable to keep up.
âThe trend to be aware of is that the number of parametersâcall this the complexity of the model,â Rao said. âThe number of parameters in a neural network model is actually increasing on the order of 10x year-on-year. This is an exponential that Iâve never seen before,â Roa noted during Intelâs most recent AI summit.
âAI is driving the compute needs of the future,â Intelâs Koduri added in announcing the SigOpt deal. âIt is even more important for software to automatically extract the best compute performance while scaling AI models.â
Artificial intelligence, Intel, Machine learning
World news – THAT – Intel Acquires Model Optimizer SigOpt