Meta offers a glimpse through its supposed iPhone killer: Orion

Image Credits: David Paul Morris/Bloomberg via Getty Images / Getty Images

For years, Silicon Valley and Wall Street have questioned Mark Zuckerberg’s decision to invest tens of billions of dollars into Reality Labs. This week, Meta’s wearables division unveiled a prototype of its Orion smart glasses, a form factor the company believes one day could replace the iPhone. That idea sounds crazy… but maybe a little less crazy than it did a week ago.

Orion is a prototype headset that combines augmented reality, eye and hand tracking, generative AI, and a gesture-detecting wristband. Through micro LED projectors and silicon carbide lenses (which are quite expensive), Meta seems to have cracked a longstanding AR display challenge. The idea is that you can look through Orion — you know, like a pair of glasses — but also see application windows projected on the lenses that appear as if they’re embedded in the world around you. Ideally, you can use your hands, eyes, and voice to navigate the environment.

Meta Orion
The Orion smart glasses need a wristband and wireless compute puck to work. (Meta)
Image Credits: Meta

Though to be clear, Meta’s Orion smart glasses are chunkier than your average readers, reportedly cost $10,000 a pop, and won’t be available for sale anytime soon. We’re talking years from now. All the technology in Orion is relatively young, and all of it needs to get cheaper, better, and smaller to work its way into a pair of smart glasses you can buy at the mall. Zuckerberg says the company has already been working on Orion for 10 years, but there’s still no path to a sellable product.

However, Meta is hardly the only company trying to put a smartphone replacement on your face.

This month, Snap unveiled its latest generation of Spectacles smart glasses, which are larger than Orion and have a more limited field of view. One former Snap engineer called the latest Spectacles “obviously bad” — though you can actually order them. Google hinted during its I/O conference in May that it, too, is working on a pair of smart glasses, perhaps a revamp of its failed Google Glass experiment from last decade. Apple is reportedly working on AR glasses that sound a lot like Orion. And we can’t rule out Jony Ive’s new startup, LoveFrom, which he recently confirmed is working on an AI wearable with OpenAI (though we don’t know if they’re glasses, a pin, or something else entirely).

What’s brewing is a race among Big Tech’s richest companies to create a sleek pair of smart glasses that can do everything your smartphone can — and hopefully something more. Meta’s prototype made two things clear: there is something there, but we’re not “there” yet.

These devices are a notable departure from the Quest virtual reality headsets Meta has been pushing for years now, and Apple’s Vision Pro. There’s a lot of similar technology involved, like eye-tracking and hand tracking, but they feel completely different to use. VR headsets are bulky, uncomfortable to wear, and make people nauseous from staring at the displays. Sunglasses and eyeglasses, on the other hand, are relatively pleasant to wear and millions of Americans use them everyday.

To Zuckerberg’s credit, he’s been pushing the eyewear form factor for quite a long time, when it certainly was not popular to do so. It’s long been reported that Meta’s CEO hates that his popular social media apps have to be accessed through Apple’s phones (perhaps leading to the ill-fated Facebook Phone). Now, Meta’s competitors are also dipping their toes into eyewear computing.

Andrew Bosworth, CTO of Meta and head or Reality Labs, wearing a clear pair of Orion smart glasses. (David Paul Morris/Bloomberg via Getty Images)

Meta’s early investment here seems to be paying off. Zuckerberg gave a keynote presentation of Orion on Wednesday that we won’t be forgetting anytime soon, filling a room full of skeptical journalists with electricity and excitement. TechCrunch has not demoed Orion yet, but initial reviews have been very positive.

What Meta offers today is the Ray-Ban Meta: a pair of glasses with cameras, microphones, speakers, sensors, an on-device LLM, and the ability to connect to your phone and the cloud. The Ray-Ban Meta is far simpler than Orion, but relatively affordable at $299 — actually not much more than a regular pair of Ray-Bans. They’re kind of like the Spectacles 3 that Snap released a few years ago, though the Ray-Ban Meta glasses appear more popular.

Despite the vast differences in price and capabilities, Orion and Ray-Ban Meta are more related than you might think.

“Orion is really the future, and we ultimately want to go for the full holographic experience. You can think about Ray-Ban Meta as our first step there,” said Li-Chen Miller, a VP of product at Meta who leads its wearables team, in an interview with TechCrunch. “We really need to nail the basic things, like making sure it’s comfortable, people want to wear it, and that people find value in it every day.”

One of the things Meta is trying to nail with Ray-Ban Meta is AI. Currently, the smart glasses use Meta’s Llama models to answer questions about what you see in front of you, by taking pictures and running them through the AI system alongside a user’s verbal requests. The Ray-Ban Meta’s AI features today are far from perfect: The latency is worse than OpenAI’s natural-feeling Advanced Voice Mode; Meta AI requires very specific prompts to work right; it hallucinates; and it doesn’t have a tight integration with many apps, making it less useful than just picking up my iPhone (perhaps by Apple’s deisgn). But Meta’s updates coming later this year try to address these issues.

Li-Chen Miller, VP of product during Meta Connect in 2023. (David Paul Morris/Bloomberg via Getty Images)

Meta announced it will soon release live AI video processing for their Ray-Bans, meaning the smart glasses will stream live video and verbal requests into one of Llama’s multimodal AI models and will produce real-time, verbal answers based on that input. It’s also getting basic features, like reminders, as well as more app integrations. That should make the whole experience a lot smoother, if it works. Miller says these improvements will filter up to Orion, which runs on the same generative AI systems.

“Some things make more sense for one form factor than the other, but we’re certainly cross-pollinating,” said Miller.

Likewise, she says some of Orion’s features may filter down as her team focuses on making the AR glasses more affordable. Orion’s various sensors and eye trackers are not cheap technologies. The problem is that Orion has to get both better and more economical.

Another challenge is typing. Your smartphone has a keyboard, but your smart glasses won’t. Miller worked on keyboards at Microsoft for nearly 20 years before joining Meta, but she says Orion’s lack of keyboard is “freeing.” She argues that using smart glasses will be a more natural experience than using a phone. You can simply talk, gesture with your hands, and look at things to navigate Orion; all things that come naturally to most people.

Another device that was criticized for lacking a keyboard was, ironically, the iPhone. Former Microsoft CEO Steve Ballmer infamously laughed at the iPhone in 2007, saying it wouldn’t appeal to business customers because it didn’t have a physical keyboard. People adapted though, and his comments sound naive more than 15 years later.

I think making Orion feel natural is definitely more of a goal than a reality at this point. The Verge notes in its hands-on review that windows occasionally filled the entire glasses lens, completely obstructing the user’s view of the world around them. That’s far from natural. To get there, Meta will have to improve its AI, typing, AR, and a long list of other features.

“For Ray-Ban Meta, we kept it very scoped to a few things, and then it does them really well,” said Miller. “Whereas, when you want to build a new, futuristic computing platform [with Orion], we have to do a lot of things, and do them all very well.”

Alexa co-creator gives first glimpse of Unlikely AI's tech strategy

William Tunstall-Pedoe, founder, Unlikely AI [© 2017 Yolande De Vries]

Image Credits: William Tunstall-Pedoe, founder, Unlikely AI (© 2017 Yolande De Vries)

After announcing a whopping $20 million seed last year, Unlikely AI founder William Tunstall-Pedoe has kept the budding U.K. foundation model maker’s approach under lock and key. Until now: TechCrunch can exclusively reveal Unlikely is taking a “neuro-symbolic” approach to its AI. In an additional development, it’s announcing two senior hires — including the former CTO of Stability AI, Tom Mason. 

Neuro-symbolic AI is a type of artificial intelligence that, as the name suggests, integrates both the modern neural network approaches to AI — as used by large language models (LLMs), like OpenAI’s GPT — and earlier Symbolic AI architectures to address the weaknesses of each.

Tunstall-Pedoe gained public profile in the U.K. tech scene back in 2012 when Amazon acquired his voice assistant startup, Evi. Two years later Amazon launched the Echo and Alexa, incorporating much of Evi’s technology. With Unlikely AI, Tunstall-Pedoe is aiming to put himself back in the limelight as he takes the wraps off the technology he and his team have been working on since 2019, when the startup was founded. 

At Stability AI, meanwhile, Mason managed the development of major foundational models across various fields and helped the AI company raise more than $170 million. Now he’s CTO of Unlikely AI, where he will oversee its “symbolic/algorithmic” approach.

In addition, Fred Becker is joining as chief administrative officer. He previously held senior roles at companies including Skype and Symphony. At Unlikely, his role will be to shepherd its now 60 full-time staff — who are based largely between Cambridge (U.K.) and London. 

The AI startup claims its approach to foundational AI models will try to avoid the risks we’ve quickly become all-too-familiar with — namely bias, “hallucination” (aka fabrication), accuracy and trust. It also claims its approach will use less energy in a bid to reduce the environmental impact of Big AI. 

“We’ve been working privately for a number of years and we’re very excited about our two new senior hires,” Tunstall-Pedoe told TechCrunch over a call. 

Fleshing out the team’s approach, he went on: “We’re building a ‘trustworthy’ AI platform that’s designed to address pretty much all of the key issues with AI at the moment, as it pertains to… hallucinations and accuracy. We’re combining the capabilities of generative AI, statistical AI, with symbolic algorithmic methods, [and] conventional software methods to get expandability and reliability.”

He described the platform as “horizontal” in that it would “compound many different types of applications.” 

Of the exact applications, he was more coy — but continued to emphasize the phrase “trustworthy AI.”

For his part, Mason said his time at Stability AI saw the company build “some amazing models” and “an unbelievable ecosystem around the models and the technology,” as he put it. It also featured the abrupt exit of founder Emad Mostaque, followed by a number of other high-profile team departures. While Mason wishes his former colleagues “all the best,” he said said he’s “super excited” to join Unlikely AI.

Stability AI CEO resigns because you’re ‘not going to beat centralized AI with more centralized AI’

Digging into the startup’s technology, Tunstall-Pedoe said the platform is composed of two things: “The word ‘neuro’ and the word ‘symbolic.’ ‘Neuro’ implies deep learning, so solving problems that machines have not been able to solve for decades… ‘Symbolic’ refers to the kind of software that powers your spreadsheets or other applications.

“One of the weaknesses of ‘neuro’ is that it’s sometimes wrong. When you train a model, you give it data, it gets better and better. But it never gets to 100%. It’s right, for example, 80% of the time, which means it’s wrong 20% of the time.”

He said this is “incredibly damaging to trust” because “the neuro calculation is opaque.” Indeed, there’s an entire field of research trying to understand what happens inside these huge LLMs.

Instead, he said Unlikely plans to combine the certainties of traditional software, such as spreadsheets, where the calculations are 100% accurate, with the “neuro” approach in generative AI. 

“What we’re doing is combining the best of both worlds,” suggested Tunstall-Pedoe. “We’re taking the capabilities of LLMs, of all the advances in deep learning, and we’re combining it with the trustworthiness and expandability and other advantages — including things like cost and environmental impact — of non-statistical machine learning… The vision we have of AI is all of those capabilities, but in a way that’s completely trustworthy.”

He argues a combined approach will bring cost and environmental benefits, too, compared to today’s LLMs: “These models are incredibly expensive [to run] and environmentally unfriendly, but they are also costly in terms of trust by producing answers that are wrong.”

Why haven’t other foundational models taken a similar route?

“I think that that’s happening,” Mason responded. “Sometimes we talk about it as ‘compound architecture.’ We’ve seen the rise of things like RAG. That’s a kind of compound architecture. This is very much in the same vein, but it’s building on all of that with the advantages symbolic reasoning, making it possible to have completely accurate reasoning.”

In this respect, he said Mason believes Unlikely AI is “ahead of the wave.”

Another question is whether Unlikely AI will produce a fuller foundational model, such as OpenAI — or take a mixed approach, akin to Mistral’s, offering both foundational and open source models?

Tunstall-Pedoe said the company is still yet to decide the direction of travel: “We haven’t made any decisions like that yet. That’s part of internal discussions. But we’re building a platform and the rest is TBD… It’s a decision that we’re going to make in the near future.”

One thing is confirmed, though: It’s going to be built out of London and Cambridge: “Obviously we’ve got a much smaller population than in the U.S. and China. But London is a fantastic place to be building an innovative AI startup. There’s lots of talent here. Lots of innovation.”

While the model release timeline isn’t clear, Unlikely AI is certain about the strength of its ambition. Given AI is the number one strategic priority of every trillion-dollar market cap company out there, Tunstall-Pedoe said he’s shooting for major adoption. “We want to be massively successful, we want to have a huge impact. We’re certainly open to different ways of achieving that,” he added.

Alexa co-creator gives first glimpse of Unlikely AI's tech strategy

William Tunstall-Pedoe, founder, Unlikely AI [© 2017 Yolande De Vries]

Image Credits: William Tunstall-Pedoe, founder, Unlikely AI (© 2017 Yolande De Vries)

After announcing a whopping $20 million seed last year, Unlikely AI founder William Tunstall-Pedoe has kept the budding U.K. foundation model maker’s approach under lock and key. Until now: TechCrunch can exclusively reveal Unlikely is taking a “neuro-symbolic” approach to its AI. In an additional development, it’s announcing two senior hires — including the former CTO of Stability AI, Tom Mason. 

Neuro-symbolic AI is a type of artificial intelligence that, as the name suggests, integrates both the modern neural network approaches to AI — as used by large language models (LLMs), like OpenAI’s GPT — and earlier Symbolic AI architectures to address the weaknesses of each.

Tunstall-Pedoe gained public profile in the U.K. tech scene back in 2012 when Amazon acquired his voice assistant startup, Evi. Two years later Amazon launched the Echo and Alexa, incorporating much of Evi’s technology. With Unlikely AI, Tunstall-Pedoe is aiming to put himself back in the limelight as he takes the wraps off the technology he and his team have been working on since 2019, when the startup was founded. 

At Stability AI, meanwhile, Mason managed the development of major foundational models across various fields and helped the AI company raise more than $170 million. Now he’s CTO of Unlikely AI, where he will oversee its “symbolic/algorithmic” approach.

In addition, Fred Becker is joining as chief administrative officer. He previously held senior roles at companies including Skype and Symphony. At Unlikely, his role will be to shepherd its now 60 full-time staff — who are based largely between Cambridge (U.K.) and London. 

The AI startup claims its approach to foundational AI models will try to avoid the risks we’ve quickly become all-too-familiar with — namely bias, “hallucination” (aka fabrication), accuracy and trust. It also claims its approach will use less energy in a bid to reduce the environmental impact of Big AI. 

“We’ve been working privately for a number of years and we’re very excited about our two new senior hires,” Tunstall-Pedoe told TechCrunch over a call. 

Fleshing out the team’s approach, he went on: “We’re building a ‘trustworthy’ AI platform that’s designed to address pretty much all of the key issues with AI at the moment, as it pertains to… hallucinations and accuracy. We’re combining the capabilities of generative AI, statistical AI, with symbolic algorithmic methods, [and] conventional software methods to get expandability and reliability.”

He described the platform as “horizontal” in that it would “compound many different types of applications.” 

Of the exact applications, he was more coy — but continued to emphasize the phrase “trustworthy AI.”

For his part, Mason said his time at Stability AI saw the company build “some amazing models” and “an unbelievable ecosystem around the models and the technology,” as he put it. It also featured the abrupt exit of founder Emad Mostaque, followed by a number of other high-profile team departures. While Mason wishes his former colleagues “all the best,” he said said he’s “super excited” to join Unlikely AI.

Stability AI CEO resigns because you’re ‘not going to beat centralized AI with more centralized AI’

Digging into the startup’s technology, Tunstall-Pedoe said the platform is composed of two things: “The word ‘neuro’ and the word ‘symbolic.’ ‘Neuro’ implies deep learning, so solving problems that machines have not been able to solve for decades… ‘Symbolic’ refers to the kind of software that powers your spreadsheets or other applications.

“One of the weaknesses of ‘neuro’ is that it’s sometimes wrong. When you train a model, you give it data, it gets better and better. But it never gets to 100%. It’s right, for example, 80% of the time, which means it’s wrong 20% of the time.”

He said this is “incredibly damaging to trust” because “the neuro calculation is opaque.” Indeed, there’s an entire field of research trying to understand what happens inside these huge LLMs.

Instead, he said Unlikely plans to combine the certainties of traditional software, such as spreadsheets, where the calculations are 100% accurate, with the “neuro” approach in generative AI. 

“What we’re doing is combining the best of both worlds,” suggested Tunstall-Pedoe. “We’re taking the capabilities of LLMs, of all the advances in deep learning, and we’re combining it with the trustworthiness and expandability and other advantages — including things like cost and environmental impact — of non-statistical machine learning… The vision we have of AI is all of those capabilities, but in a way that’s completely trustworthy.”

He argues a combined approach will bring cost and environmental benefits, too, compared to today’s LLMs: “These models are incredibly expensive [to run] and environmentally unfriendly, but they are also costly in terms of trust by producing answers that are wrong.”

Why haven’t other foundational models taken a similar route?

“I think that that’s happening,” Mason responded. “Sometimes we talk about it as ‘compound architecture.’ We’ve seen the rise of things like RAG. That’s a kind of compound architecture. This is very much in the same vein, but it’s building on all of that with the advantages symbolic reasoning, making it possible to have completely accurate reasoning.”

In this respect, he said Mason believes Unlikely AI is “ahead of the wave.”

Another question is whether Unlikely AI will produce a fuller foundational model, such as OpenAI — or take a mixed approach, akin to Mistral’s, offering both foundational and open source models?

Tunstall-Pedoe said the company is still yet to decide the direction of travel: “We haven’t made any decisions like that yet. That’s part of internal discussions. But we’re building a platform and the rest is TBD… It’s a decision that we’re going to make in the near future.”

One thing is confirmed, though: It’s going to be built out of London and Cambridge: “Obviously we’ve got a much smaller population than in the U.S. and China. But London is a fantastic place to be building an innovative AI startup. There’s lots of talent here. Lots of innovation.”

While the model release timeline isn’t clear, Unlikely AI is certain about the strength of its ambition. Given AI is the number one strategic priority of every trillion-dollar market cap company out there, Tunstall-Pedoe said he’s shooting for major adoption. “We want to be massively successful, we want to have a huge impact. We’re certainly open to different ways of achieving that,” he added.

Alexa co-creator gives first glimpse of Unlikely AI's tech strategy

William Tunstall-Pedoe, founder, Unlikely AI [© 2017 Yolande De Vries]

Image Credits: William Tunstall-Pedoe, founder, Unlikely AI (© 2017 Yolande De Vries)

After announcing a whopping $20 million seed last year, Unlikely AI founder William Tunstall-Pedoe has kept the budding U.K. foundation model maker’s approach under lock and key. Until now: TechCrunch can exclusively reveal Unlikely is taking a “neuro-symbolic” approach to its AI. In an additional development, it’s announcing two senior hires — including the former CTO of Stability AI, Tom Mason. 

Neuro-symbolic AI is a type of artificial intelligence that, as the name suggests, integrates both the modern neural network approaches to AI — as used by large language models (LLMs), like OpenAI’s GPT — and earlier Symbolic AI architectures to address the weaknesses of each.

Tunstall-Pedoe gained public profile in the U.K. tech scene back in 2012 when Amazon acquired his voice assistant startup, Evi. Two years later Amazon launched the Echo and Alexa, incorporating much of Evi’s technology. With Unlikely AI, Tunstall-Pedoe is aiming to put himself back in the limelight as he takes the wraps off the technology he and his team have been working on since 2019, when the startup was founded. 

At Stability AI, meanwhile, Mason managed the development of major foundational models across various fields and helped the AI company raise more than $170 million. Now he’s CTO of Unlikely AI, where he will oversee its “symbolic/algorithmic” approach.

In addition, Fred Becker is joining as chief administrative officer. He previously held senior roles at companies including Skype and Symphony. At Unlikely, his role will be to shepherd its now 60 full-time staff — who are based largely between Cambridge (U.K.) and London. 

The AI startup claims its approach to foundational AI models will try to avoid the risks we’ve quickly become all-too-familiar with — namely bias, “hallucination” (aka fabrication), accuracy and trust. It also claims its approach will use less energy in a bid to reduce the environmental impact of Big AI. 

“We’ve been working privately for a number of years and we’re very excited about our two new senior hires,” Tunstall-Pedoe told TechCrunch over a call. 

Fleshing out the team’s approach, he went on: “We’re building a ‘trustworthy’ AI platform that’s designed to address pretty much all of the key issues with AI at the moment, as it pertains to… hallucinations and accuracy. We’re combining the capabilities of generative AI, statistical AI, with symbolic algorithmic methods, [and] conventional software methods to get expandability and reliability.”

He described the platform as “horizontal” in that it would “compound many different types of applications.” 

Of the exact applications, he was more coy — but continued to emphasize the phrase “trustworthy AI.”

For his part, Mason said his time at Stability AI saw the company build “some amazing models” and “an unbelievable ecosystem around the models and the technology,” as he put it. It also featured the abrupt exit of founder Emad Mostaque, followed by a number of other high-profile team departures. While Mason wishes his former colleagues “all the best,” he said said he’s “super excited” to join Unlikely AI.

Stability AI CEO resigns because you’re ‘not going to beat centralized AI with more centralized AI’

Digging into the startup’s technology, Tunstall-Pedoe said the platform is composed of two things: “The word ‘neuro’ and the word ‘symbolic.’ ‘Neuro’ implies deep learning, so solving problems that machines have not been able to solve for decades… ‘Symbolic’ refers to the kind of software that powers your spreadsheets or other applications.

“One of the weaknesses of ‘neuro’ is that it’s sometimes wrong. When you train a model, you give it data, it gets better and better. But it never gets to 100%. It’s right, for example, 80% of the time, which means it’s wrong 20% of the time.”

He said this is “incredibly damaging to trust” because “the neuro calculation is opaque.” Indeed, there’s an entire field of research trying to understand what happens inside these huge LLMs.

Instead, he said Unlikely plans to combine the certainties of traditional software, such as spreadsheets, where the calculations are 100% accurate, with the “neuro” approach in generative AI. 

“What we’re doing is combining the best of both worlds,” suggested Tunstall-Pedoe. “We’re taking the capabilities of LLMs, of all the advances in deep learning, and we’re combining it with the trustworthiness and expandability and other advantages — including things like cost and environmental impact — of non-statistical machine learning… The vision we have of AI is all of those capabilities, but in a way that’s completely trustworthy.”

He argues a combined approach will bring cost and environmental benefits, too, compared to today’s LLMs: “These models are incredibly expensive [to run] and environmentally unfriendly, but they are also costly in terms of trust by producing answers that are wrong.”

Why haven’t other foundational models taken a similar route?

“I think that that’s happening,” Mason responded. “Sometimes we talk about it as ‘compound architecture.’ We’ve seen the rise of things like RAG. That’s a kind of compound architecture. This is very much in the same vein, but it’s building on all of that with the advantages symbolic reasoning, making it possible to have completely accurate reasoning.”

In this respect, he said Mason believes Unlikely AI is “ahead of the wave.”

Another question is whether Unlikely AI will produce a fuller foundational model, such as OpenAI — or take a mixed approach, akin to Mistral’s, offering both foundational and open source models?

Tunstall-Pedoe said the company is still yet to decide the direction of travel: “We haven’t made any decisions like that yet. That’s part of internal discussions. But we’re building a platform and the rest is TBD… It’s a decision that we’re going to make in the near future.”

One thing is confirmed, though: It’s going to be built out of London and Cambridge: “Obviously we’ve got a much smaller population than in the U.S. and China. But London is a fantastic place to be building an innovative AI startup. There’s lots of talent here. Lots of innovation.”

While the model release timeline isn’t clear, Unlikely AI is certain about the strength of its ambition. Given AI is the number one strategic priority of every trillion-dollar market cap company out there, Tunstall-Pedoe said he’s shooting for major adoption. “We want to be massively successful, we want to have a huge impact. We’re certainly open to different ways of achieving that,” he added.

Close up of hands typing code on a keyboard with code appearing on monitor in front of the keyboard.

Alibaba staffer offers a glimpse into building LLMs in China

Close up of hands typing code on a keyboard with code appearing on monitor in front of the keyboard.

Image Credits: gorodenkoff / Getty Images

Chinese tech companies are gathering all sorts of resources and talent to narrow their gap with OpenAI, and experiences for researchers on both sides of the Pacific Ocean can be surprisingly similar. A recent X post from an Alibaba researcher offers a rare glimpse into the life of developing large language models at the e-commerce firm, which is among a raft of Chinese internet giants striving to match the capabilities of ChatGPT.

Binyuan Hui, a natural language processing researcher at Alibaba’s large language model team Qwen, shared his daily schedule on X, mirroring a post by OpenAI researcher Jason Wei that went viral recently.

The parallel glimpse into their typical day reveals striking similarities, with wake-up times at 9 a.m. and bedtime around 1 a.m. Both start the day with meetings, followed by a period of coding, model training and brainstorming with colleagues. Even after getting home, they continue to run experiments at night and ponder on ways to enhance their models well into bedtime.

The notable differences are in how they choose to characterize leisure time. Hui, the Alibaba employee, mentioned reading research papers and browsing X to catch up on “what is happening in the world.” And as a commentator pointed out, Hui doesn’t have a glass of wine after he arrives home like Wei does.

This intense work regime is not unusual in China’s current LLM space, where tech talent with top university degrees are joining tech companies in droves to build competitive AI models.

To a certain extent, Hui’s demanding schedule seems to reflect a personal drive to match (or at least the social media appearance of doing so), if not outpace, Silicon Valley companies in the AI space. It seems different from the involuntary “996” work hours associated with more “traditional” types of Chinese internet businesses that involve heavy operations, such as video games and e-commerce.

Indeed, even renowned AI investor and computer scientist Kai-Fu Lee puts in an incredible amount of effort. When I interviewed Lee about his newly minted LLM unicorn 01.AI in November, he admitted that late hours were the norm, but employees were willingly working hard. That day, one of his staff messaged him at 2:15 a.m. to express his excitement about being part of 01.AI’s mission.

Outward displays of intense work ethic speak to the urgency of the remits laid out by tech firms in the country, and subsequently the speed with which those firms are now rolling out LLMs.

Qwen, for example, has open sourced a series of foundation models trained with both English and Chinese data. The number of parameters — a figure that speaks to the knowledge the model gains from historical training data that defines its ability to generate contextually relevant responses — is 72 billion for the largest of these. (For some context, GPT3 from OpenAI is believed to have 175 billion; GPT4, its latest LLM, has 1.7 trillion. However, it’s arguable that the aim of a particular LLM will be the more important key to decoding the value of high parameter numbers.)

The team also has been quick to introduce commercial applications. Last April, Alibaba began integrating Qwen into its enterprise communication platform DingTalk and online retailer Tmall.

No definite leader has emerged in China’s LLM space so far, and venture capital firms and corporate investors are spreading their bets across multiple contenders. Besides building its own LLM in-house, Alibaba has been aggressively investing in startups such as Moonshot AI, Zhipu AI, Baichuan and 01.AI.

Facing competition, Alibaba has been trying to carve out a niche, and its multilingual move could become a selling point. In December, the company released an LLM for several Southeast Asian languages. Called SeaLLM, the model is capable of processing information in Vietnamese, Indonesian, Thai, Malay, Khmer, Lao, Tagalog and Burmese. Through its cloud computing business and acquisition of e-commerce platform Lazada, Alibaba has established a sizable footprint in the region and can potentially introduce SeaLLM to these services down the road.

How China is building a parallel generative AI universe