Meta Platforms has hired an Oslo-based team that until late last year was building artificial intelligence networking technology at British chip unicorn Graphcore.
A Meta spokesman confirmed the hirings in response to a request for comment, after Reuters identified 10 people whose LinkedIn profiles said they worked at Graphcore until December 2022 or January 2023 and subsequently joined Meta in February or March of this year.
“We recently welcomed a number of highly specialised engineers in Oslo to our infrastructure team at Meta. They bring deep expertise in the design and development of supercomputing systems to support AI and machine learning at scale in Meta’s data centres,” said Jon Carvill, the Meta spokesman.
The move brings additional muscle to the social media giant’s bid to improve how its data centres handle AI work, as it races to cope with demand for AI-orientated infrastructure from teams across the company looking to build new features.
Meta, which owns Facebook and Instagram, has become increasingly reliant on AI technology to target advertising, select posts for its apps’ feeds and purge banned content from its platforms.
On top of that, it is now rushing to join competitors such as Microsoft and Google in releasing generative AI products capable of creating human-like writing, art and other content, which investors see as the next big growth area for tech companies.
The 10 employees’ job descriptions on LinkedIn indicated the team had worked on AI-specific networking technology at Graphcore, which develops computer chips and systems optimised for AI work.
Carvill declined to say what they would be working on at Meta.
Graphcore closed its Oslo office as part of a broader restructuring announced in October last year, a spokesman for the start-up said, as it struggled to make inroads against US-based firms such as Nvidia and AMD, which dominate the market for AI chips.
Meta already has an in-house unit designing several kinds of chips aimed at speeding up and maximising efficiency for its AI work, including a network chip that performs a sort of air traffic control function for servers, two sources said.
Efficient networking is especially useful for modern AI systems like those behind chatbot ChatGPT or image-generation tool Dall-E, which are far too large to fit onto a single computing chip and must instead be split up over many chips strung together.
A new category of network chip has emerged to help keep data moving smoothly within those computing clusters. Nvidia, AMD and Intel all make such network chips.
In addition to its network chip, Meta is also designing a complex computing chip to both train AI models and perform inference, a process in which the trained models make judgments and generate responses to prompts, although it does not expect that chip to be ready until around 2025.
Graphcore, one of the UK’s most valuable tech start-ups, once was seen by investors like Microsoft and venture capital firm Sequoia as a promising potential challenger to Nvidia’s commanding lead in the market for AI chip systems.
However, it faced a setback in 2020 when Microsoft scrapped an early deal to buy Graphcore’s chips for its Azure cloud computing platform, according to a report by UK newspaper The Times. Microsoft instead used Nvidia’s GPUs to build the massive infrastructure powering ChatGPT developer OpenAI, which Microsoft also backs.
Sequoia has since written down its investment in Graphcore to zero, although it remains on the company’s board, according to a source familiar with the relationship.
The Graphcore spokesman confirmed the setbacks, but said the company was “perfectly positioned” to take advantage of accelerating commercial adoption of AI.
Graphcore was last valued at US$2.8-billion after raising $222-million in its most recent investment round in 2020. — Katie Paul, Krystal Hu and Stephen Nellis, (c) 2023 Reuters