Meta AI
View on WikipediaThis article may contain excessive or inappropriate references to self-published sources. (July 2025) |
| This article is part of a series about |
| Meta Platforms |
|---|
| Products and services |
| People |
| Business |
| Part of a series on |
| Artificial intelligence (AI) |
|---|
Meta AI is a research division of Meta (formerly Facebook) that develops artificial intelligence and augmented reality technologies.
Key Information
History
[edit]Meta AI was founded in 2013 as Facebook Artificial Intelligence Research (FAIR).[1][2] It has workspaces in Menlo Park, London, New York City, Paris, Seattle, Pittsburgh, Tel Aviv, and Montreal as of 2025.[3][4]
In 2016, FAIR partnered with Google, Amazon, IBM, and Microsoft in creating the Partnership on Artificial Intelligence to Benefit People and Society.
Meta AI was directed by Yann LeCun until 2018, when Jérôme Pesenti succeeded the role. Pesenti is formerly the CTO of IBM's big data group.[5]
FAIR's research includes self-supervised learning, generative adversarial networks, document classification and translation, and computer vision.[6] FAIR released Torch deep-learning modules as well as PyTorch in 2017, an open-source machine learning framework,[6] which was subsequently used in several deep learning technologies, such as Tesla's autopilot [7] and Uber's Pyro.[8] That same year, a pair of chatbots were falsely rumored[9] to be discontinued for developing a language that was unintelligible to humans.[10] FAIR clarified that the research had been shut down because they had accomplished their initial goal to understand how languages are generated by their models, rather than out of fear.[9]
FAIR was renamed Meta AI following the rebranding that changed Facebook, Inc. to Meta Platforms Inc.[11]
Virtual assistant
[edit]Meta AI is also the name of the virtual assistant developed by the team, now integrated as a chatbot into Meta's social networking products.[12] It is also available as a subscription-based stand-alone app.[13][14]
The virtual assistant was pre-installed on the second generation of Ray-Ban Meta smartglasses, and can incorporate inputs from the glasses' cameras after an update.[15] It is also available on Quest 2 and newer HMDs.[16]
Since May 2024, the chatbot has summarized news from various outlets without linking directly to original articles, including in Canada, where news links are banned on its platforms. This use of news content without compensation and attribution has raised ethical and legal concerns, especially as Meta continues to reduce news visibility on its platforms.[17]
Current research
[edit]This section possibly contains original research. cited articles are all original researches instead of e.g., reviews. (July 2025) |
Natural language processing and chatbot
[edit]Meta AI works on machines' ability to understand and generate natural language. The team also seeks to allow their chatbots to communicate multilingually.[18] This involves the generalization of natural language processing (NLP) technology to other languages, and the team actively works on unsupervised machine translation.[19][20]
Galactica
[edit]Galactica is a large language model (LLM) designed for generating scientific text. It was available for three days from 15 November 2022, before being withdrawn for generating racist and inaccurate content.[21][22]
Llama
[edit]LLaMA is a LLM released in February 2023, supporting 7B to 65B parameters.[23] Two of the three Llama 4 models, Scout and Maverick, were released on April 5, 2025, with the biggest model, Behemoth, still in training.[24]
Hardware
[edit]Meta used CPUs and in-house custom chips until 2022, when they switched to Nvidia GPUs. Several data centers were redesigned to accommodate the larger network bandwidth and cooling requirements.[25]
MTIA v1
[edit]Meta developed the training and inference accelerator, MTIA v1, specifically for their content recommendation workloads. It was fabricated on TSMC's 7 nm process technology and operates at a frequency of 800 MHz. The accelerator provides 51.2 TFLOPS at FP16 precision, with a thermal design power (TDP) of 25 W.[26]
References
[edit]- ^ "NYU "Deep Learning" Professor LeCun Will Head Facebook's New Artificial Intelligence Lab". TechCrunch. 9 December 2013. Retrieved 2022-05-08.
- ^ "Facebook's AI team hires Vladimir Vapnik, father of the popular support vector machine algorithm". VentureBeat. 2014-11-25. Archived from the original on 2014-11-27. Retrieved 2022-05-08.
- ^ "Facebook Opens New AI Research Center In Paris". TechCrunch. 2 June 2015. Retrieved 2022-05-08.
- ^ Dillet, Romain (June 2, 2015). "Facebook Opens New AI Research Center in Paris". TechCrunch. Retrieved May 7, 2022.
- ^ Dave, Greshgorn (January 23, 2018). "The head of Facebook's AI research is stepping into a new role as it shakes up management". Quartz. Archived from the original on May 8, 2022. Retrieved May 7, 2022.
- ^ a b "FAIR turns five: What we've accomplished and where we're headed". Engineering at Meta. 2018-12-05. Archived from the original on 2022-05-11. Retrieved 2022-05-08.
- ^ Karpathy, Andrej (6 November 2019). "PyTorch at Tesla - Andrej Karpathy, Tesla". YouTube. Archived from the original on 2023-03-24. Retrieved 2022-05-08.
- ^ "Pyro". pyro.ai. Archived from the original on 2022-05-06. Retrieved 2022-05-08.
- ^ a b "Facebook researchers shut down AI bots that started speaking in a language unintelligible to humans". Tech2. 2017-07-31. Archived from the original on 2022-05-08. Retrieved 2022-05-08.
- ^ McKay, Tom (2017-08-01). "No, Facebook Did Not Panic and Shut Down an AI Program That Was Getting Dangerously Smart". Gizmodo. Retrieved 2025-05-27.
When Facebook directed two of these semi-intelligent bots to talk to each other, FastCo reported, the programmers realized they had made an error by not incentivizing the chatbots to communicate according to human-comprehensible rules of the English language. In their attempts to learn from each other, the bots thus began chatting back and forth in a derived shorthand—but while it might look creepy, that's all it was.
- ^ Murphy Kelly, Samantha (October 29, 2021). "Facebook changes its company name to Meta". CNN Business. Archived from the original on May 7, 2022. Retrieved May 7, 2022.
- ^ Kawale, Ajinkya (20 February 2025). "India among largest Meta AI adopters, backs open-source innovation". Business Standard. Archived from the original on 2025-02-20. Retrieved 2025-02-22.
- ^ "Meta to Launch Standalone AI App with Premium Features Amid Growing Competition". Mint. 2024-02-27. Retrieved 2025-02-28.
- ^ "Meta AI Expansion: Standalone App and Subscription Model in the Works". Mint. 2024-02-27. Retrieved 2025-02-28.
- ^ "Smart(er) Glasses: Introducing New Ray-Ban | Meta Styles + Expanding Access to Meta AI with Vision". Meta Quest Blog. 2024-04-23. Archived from the original on 2024-07-27.
- ^ Meta Quest Blog (July 23, 2024). "Introducing Meta AI on Meta Quest—Your Smart MR Assistant". Meta Blog.
- ^ "Meta walked away from news. Now the company's using it for AI content". The Washington Post. 21 May 2024. Archived from the original on 21 May 2024. Retrieved 22 May 2024.
- ^ "Meta AI Research Topic - Natural Language Processing". ai.facebook.com. Archived from the original on 2022-05-08. Retrieved 2022-05-08.
- ^ Lample, Guillaume; Ott, Myle; Conneau, Alexis; Denoyer, Ludovic; Ranzato, Marc'Aurelio (2018-08-13). "Phrase-Based & Neural Unsupervised Machine Translation". arXiv:1804.07755 [cs.CL].
- ^ Conneau, Alexis; Lample, Guillaume; Rinott, Ruty; Williams, Adina; Bowman, Samuel R.; Schwenk, Holger; Stoyanov, Veselin (2018-09-13). "XNLI: Evaluating Cross-lingual Sentence Representations". arXiv:1809.05053 [cs.CL].
- ^ "Why Meta's latest large language model survived only three days online". MIT Technology Review. Retrieved 2025-07-18.
- ^ Edwards, Benj (18 November 2022). "New Meta AI demo writes racist and inaccurate scientific literature, gets pulled". Ars Technica. Retrieved 30 December 2022.
- ^ Leswing, Kif (2023-02-24). "Mark Zuckerberg announces Meta's new large language model as A.I. race heats up". CNBC. Retrieved 2025-04-14.
- ^ Wiggers, Kyle (2025-04-05). "Meta releases Llama 4, a new crop of flagship AI models". TechCrunch. Retrieved 2025-04-14.
- ^ "Insight: Inside Meta's scramble to catch up on AI". Reuters. 2023-04-26.
- ^ Peters, Jay (2023-05-19). "Meta is working on a new chip for AI". The Verge. Archived from the original on 2023-06-07. Retrieved 2023-06-07.
External links
[edit]Meta AI
View on GrokipediaHistory
Founding and Early Development
Facebook AI Research (FAIR), the foundational entity behind Meta AI's development, was established in December 2013 by Facebook (now Meta Platforms, Inc.) to advance artificial intelligence through rigorous, open scientific inquiry.[12] The initiative stemmed from CEO Mark Zuckerberg's recognition of AI's potential to improve platform features like content recommendation and user interaction, while also pursuing broader goals of understanding human-level intelligence.[13] FAIR's charter emphasized fundamental research over immediate product applications, with a commitment to sharing findings via publications and open-source code to accelerate global progress.[14] Yann LeCun, a leading expert in machine learning and convolutional neural networks, joined as FAIR's first director that same month, recruited personally by Zuckerberg amid competition for top talent.[15] The initial team, small and New York-based, concentrated on core challenges in deep learning, computer vision, speech recognition, and reasoning systems, producing early breakthroughs such as improved object detection algorithms and contributions to large-scale neural network training.[12] LeCun's leadership prioritized long-term paradigm shifts in AI, drawing from his prior work at institutions like NYU and Bell Labs, rather than short-term engineering fixes.[16] By 2015, FAIR had grown to include an international outpost in Paris, leveraging Europe's deep expertise in mathematics and AI to bolster efforts in areas like natural language understanding and reinforcement learning.[17] This expansion enabled collaborative projects, including early experiments with multi-modal AI systems that integrated text, images, and video—precursors to later consumer tools.[13] The lab's output during this period included high-impact publications at conferences like NeurIPS and CVPR, alongside releases of datasets and toolkits that influenced the broader AI community, solidifying FAIR's role as a hub for empirical, data-driven advancements.[12]Evolution into Core Division
Facebook Artificial Intelligence Research (FAIR), established on December 9, 2013, initially operated as a dedicated lab focused on fundamental advancements in machine learning, computer vision, and natural language processing, emphasizing open-source contributions to benefit the broader AI community.[12] Early efforts prioritized exploratory research over immediate product applications, with Yann LeCun appointed as founding director to lead theoretical breakthroughs.[12] Contributions from FAIR gradually influenced Meta's operational infrastructure, notably through the development of PyTorch in 2016, an open-source deep learning framework that transitioned from a research prototype to a cornerstone for scalable AI deployment across Meta's engineering teams.[18] This enabled practical integrations, such as enhanced recommendation algorithms in feeds and targeted advertising systems, where AI had been foundational since 2006 but accelerated with FAIR's tools for handling vast datasets from billions of users.[19] The competitive pressure following OpenAI's ChatGPT release in November 2022 catalyzed a strategic escalation, with Meta reallocating resources to generative AI amid a broader pivot from metaverse priorities.[20] In February 2023, Meta unified its generative AI initiatives under a new product group, shifting focus from siloed research to rapid incorporation of technologies like large language models into consumer-facing apps, including Instagram, Facebook, and WhatsApp.[21] This reorganization marked AI's elevation from peripheral R&D to a cross-functional priority, supported by commitments to annual capital expenditures exceeding $9.5 billion for AI-specific compute infrastructure by late 2023.[22] By September 27, 2023, Meta launched its flagship Meta AI assistant, powered by Llama 2 models and integrated directly into messaging and social features, positioning AI as a core engagement driver rather than an experimental add-on.[23] CEO Mark Zuckerberg articulated this as embedding AI "into every product" to enhance personalization and utility, with generative capabilities extending to advertising tools and content creation, thereby aligning research outputs with revenue-generating functions like ad optimization, which constitutes over 97% of Meta's income.[19] Subsequent refinements, including 2025 team splits for dedicated product integration streams, reinforced this trajectory, streamlining decision-making to prioritize applied AI over pure academia-style inquiry.[24]Major Milestones and Shifts (2013–2025)
In 2013, Facebook established the Fundamental AI Research (FAIR) lab on December 9, with Yann LeCun appointed as its founding director, marking the inception of systematic AI research efforts focused on areas such as computer vision, natural language processing, and machine learning fundamentals.[12] The lab initially operated from New York and emphasized open research practices, contributing early advancements like improvements in deep learning architectures that influenced subsequent industry developments.[14] By 2016–2018, FAIR expanded globally with new labs in London, Paris, Montreal, and Pittsburgh, while achieving recognition through multiple Best Paper awards at conferences including ACL, CVPR, and ECCV, alongside Test of Time honors for prior work.[13] A pivotal output was the development and initial release of PyTorch in 2017, an open-source deep learning framework that facilitated broader adoption of dynamic neural networks and became a cornerstone for AI experimentation worldwide.[12] This period reflected a shift from isolated academic pursuits to tools enabling scalable AI deployment, though FAIR remained primarily research-oriented without direct product integration. The 2020s brought a strategic pivot toward generative AI and practical applications, accelerated by the February 2023 release of LLaMA 1, a family of efficient large language models initially available for research, which demonstrated competitive performance on benchmarks despite smaller sizes compared to proprietary rivals.[12] In July 2023, Meta open-sourced LLaMA 2, expanding access under a commercial license and powering the September 27 launch of the Meta AI assistant—a multimodal chatbot integrated into Facebook, Instagram, Messenger, and WhatsApp for tasks like content generation and query resolution.[25] This marked FAIR's evolution from pure research to consumer-facing products, with Meta AI achieving nearly 600 million monthly active users by late 2024.[26] Subsequent model iterations underscored rapid scaling: LLaMA 3 launched on April 18, 2024, with 8B and 70B parameter variants outperforming prior open models on reasoning and coding benchmarks; LLaMA 3.1 followed in July 2024, extending context length to 128,000 tokens and adding multilingual support.[25] LLaMA 3.2 introduced multimodal capabilities in December 2024, while LLaMA 4 was released by January 2026, featuring models optimized for efficiency.[26] On April 29, 2025, Meta released a standalone Meta AI app, enhancing accessibility beyond platform integrations and emphasizing personalized, context-aware interactions.[6] Amid these advancements, 2025 saw internal shifts, including the October layoff of approximately 600 roles across FAIR and related AI units, redirecting resources toward superintelligence pursuits and infrastructure investments exceeding $65 billion annually to support advanced model training.[27] Later in 2025, Meta introduced the Vibes feature in September and released SAM 3 in November, advancing computer vision capabilities.[28] Announcements at Meta Connect 2025 highlighted ongoing integrations into products like Ray-Ban smartglasses and Quest headsets. In early 2026, Llama 4 was fully released by January, followed by an AMD partnership on February 24 to enhance AI infrastructure. These developments, coupled with strong adoption in regions like India and sustained open-source efforts, reflect continued visibility and momentum in Meta AI's evolution. In December 2025, Meta acquired Manus, a Singapore-based AI agent company, for approximately $2 billion to accelerate automation integration across consumer and enterprise products; Manus had achieved over $100 million in annualized revenue within eight months of launching its general-purpose AI agent capable of market research, coding, and data analysis. Meta is also developing next-generation models codenamed "Mango" and "Avocado," targeting release in the first half of 2026. This restructuring highlighted a tension between open-source commitments and competitive pressures, as Meta balanced foundational research with proprietary enhancements for edge in reasoning and multimodality.[26]Recent Developments
In March 2026, Meta granted significant stock options to top executives as part of efforts to retain talent amid the competitive AI landscape. Concurrently, the company conducted layoffs affecting several hundred positions across Reality Labs, Facebook, and other divisions to redirect resources toward AI priorities.Organizational Structure and Leadership
Key Leaders and Roles
Yann LeCun serves as Meta's Chief AI Scientist and Vice President, a position he has held since joining the company in December 2013 to lead the Fundamental AI Research (FAIR) lab.[15] In this capacity, LeCun directs foundational research in areas such as deep learning, convolutional neural networks, and self-supervised learning, drawing on his prior work as a pioneer in these fields.[29] His leadership emphasizes long-term AI advancements over short-term product applications, as evidenced by FAIR's contributions to open-source models like Llama.[30] In June 2025, Meta established the Meta Superintelligence Labs (MSL) and appointed Alexandr Wang, the 28-year-old former CEO of Scale AI, as the company's inaugural Chief AI Officer to head the initiative.[31] MSL consolidated all AI teams into four divisions: TBD Lab for foundation model development, FAIR for fundamental research, Products and Applied Research, and MSL Infrastructure. Wang oversees MSL's efforts to build highly capable AI systems, including large-scale model training and recruitment of top talent from competitors like OpenAI and DeepMind, amid Meta's $14.3 billion investment in Scale AI.[32] This role positions him to consolidate decision-making across AI teams, as demonstrated by his oversight of a October 2025 restructuring that eliminated approximately 600 positions to streamline operations.[33] FAIR's leadership transitioned in May 2025 when Joëlle Pineau, who had served as Vice President of AI Research since 2019 and managed aspects of generative AI and reinforcement learning, departed to become Chief AI Officer at Cohere.[34] Robert Fergus, formerly a director at Google DeepMind, was appointed to lead FAIR in her place, focusing on core research continuity amid Meta's shift toward applied superintelligence pursuits.[35] Overall AI strategy remains under the purview of CEO Mark Zuckerberg, who has directed multiple reorganizations to prioritize scalable AI infrastructure.[30]Restructurings and Workforce Changes
In October 2025, Meta Platforms announced the elimination of approximately 600 positions across its artificial intelligence division, including teams within Fundamental AI Research (FAIR), product-related AI groups, and AI infrastructure units.[36][32] The cuts, detailed in an internal memo from Chief AI Officer Alexandr Wang, targeted bureaucratic layers to enable faster decision-making, more direct communication, and greater individual ownership amid intensified competition in AI development.[37][38] This restructuring affected Superintelligence Labs, a key AI initiative, but occurred alongside continued hiring for specialized roles in advanced AI labs, reflecting a selective refinement rather than broad contraction.[39][27] The layoffs followed Meta's aggressive talent acquisition earlier in 2025, including the recruitment of over 50 researchers from rival labs, which contributed to organizational bloat in non-core areas.[38] Company executives framed the changes as necessary to align workforce structure with strategic priorities, such as scaling superintelligence efforts, while maintaining heavy investments—exceeding billions annually—in AI infrastructure and compute resources.[40][32] Prior to this, Meta's AI teams had largely avoided the broader corporate layoffs of 2022 (11,000 roles) and 2023 (over 10,000 roles), as the company pivoted toward AI expansion by hiring hundreds of specialized engineers and scientists to bolster capabilities in large language models and generative technologies.[41] These adjustments underscore Meta's iterative approach to AI organization, balancing rapid scaling with efficiency drives, even as overall headcount in core AI functions remains elevated compared to pre-2022 levels.[42] No significant prior restructurings unique to the AI division were publicly detailed beyond integration of FAIR into broader Meta AI operations in 2023, which emphasized cross-platform AI deployment without reported mass workforce shifts.[43]Research Focus Areas
Fundamental AI Research (FAIR)
Meta's Fundamental AI Research (FAIR) lab is the company's primary center for long-term and fundamental artificial intelligence research. Founded in December 2013, with Yann LeCun as its founding director, FAIR was established to pursue open scientific inquiry into AI, focusing on foundational advancements rather than immediate commercial applications. Current research at FAIR encompasses self-supervised learning, world models for reasoning and planning, embodied AI, multimodal understanding, and scalable architectures for advanced intelligence. Breakthroughs include the creation of PyTorch, pioneering work in convolutional neural networks and deep learning, the LLaMA family of open language models, the Segment Anything Model (SAM) series for vision tasks, and innovations in reinforcement learning and robotics through platforms like Habitat. Led by Yann LeCun, who advocates for objective-driven AI and predictive world models (such as JEPA), FAIR continues to release open-source contributions to accelerate progress in the global AI community while supporting Meta's broader AI ecosystem.Large Language Models
Meta AI's large language models are primarily embodied in the LLaMA family, a series of transformer-based autoregressive models developed to advance natural language understanding and generation through efficient scaling and optimization. Initiated with LLaMA 1 in February 2023, featuring variants from 7 billion to 65 billion parameters trained on approximately 1.4 trillion tokens of public internet data, these models prioritized research utility and parameter efficiency over sheer scale. Early releases demonstrated competitive performance on benchmarks like GLUE and SuperGLUE, often rivaling larger proprietary systems despite smaller sizes, due to architectural refinements such as grouped-query attention and rotary positional embeddings.[44] LLaMA 2, released in July 2023, expanded to 7B, 13B, and 70B parameter models, incorporating safety alignments via supervised fine-tuning and reinforcement learning from human feedback to mitigate harmful outputs. This iteration processed over 2 trillion tokens during training, achieving scores such as 68.9% on MMLU for the 70B variant, positioning it as a benchmark for open research models. LLaMA 3 followed on April 18, 2024, with 8B and 70B pretrained and instruction-tuned versions trained on more than 15 trillion tokens, enhancing reasoning capabilities evidenced by improvements in coding tasks (e.g., 68.4% on HumanEval for 70B) and multilingual support across 30+ languages.[25]| Model Version | Release Date | Parameter Sizes | Notable Benchmarks and Features |
|---|---|---|---|
| LLaMA 3 | April 18, 2024 | 8B, 70B | MMLU: up to 82.0% (70B instruct); extended vocabulary, tool-use integration; trained on 15T+ tokens.[25] |
| LLaMA 3.1 | July 23, 2024 | 8B, 70B, 405B | MMLU: 88.6% (405B); supports 128K context, multilingual (8 languages), outperforms GPT-3.5 on 150+ evals.[5] |
| LLaMA 3.2 | September 2024 | 1B, 3B (text); 11B, 90B (vision) | Added vision-language capabilities; lightweight for edge deployment.[45] |
| LLaMA 3.3 | December 6, 2024 | 70B | Matches 405B performance on select tasks; optimized for inference efficiency.[46] |
| LLaMA 4 | April 5, 2025 | Scout (17B active/109B total), Maverick | Native multimodality (text+image); up to 1M token context; open-weight for research.[4] [47] |
Other AI Research Initiatives
Meta AI, formerly Facebook AI Research (FAIR), has made significant contributions to natural language processing (NLP) research, including numerous publications, awards, and influential models and datasets at the Association for Computational Linguistics (ACL) and related venues. Notable examples include the ACL 2018 best paper honorable mention for "Hierarchical Neural Story Generation"[48], leadership in machine translation and low-resource NLP, the development of the PyText framework for deep-learning-based NLP modeling[49], and the XNLI dataset for cross-lingual sentence representation evaluation.[50] In 2024, Scott Wen-tau Yih from Meta FAIR was elected an ACL Fellow for contributions to information extraction, question answering, neural retrieval, and retrieval-augmented generation.[51] These efforts encompass deep learning applications in machine translation, natural language understanding, dialogue systems, and cross-lingual transfer. Meta AI's computer vision research emphasizes foundational models for visual understanding and segmentation. The Segment Anything Model (SAM), released on April 5, 2023, introduced promptable segmentation capable of identifying and outlining any object in an image with minimal user input, trained on over 1 billion masks from the SA-1B dataset comprising 11 million images. Its successor, SAM 2, launched on July 30, 2024, extended capabilities to video by enabling real-time object tracking and segmentation across frames, supporting applications in video editing and augmented reality. SAM 3, released on November 19, 2025, further advanced this with a unified model supporting text, exemplar, and visual prompts for detection, segmentation, and tracking of concepts in images and videos.[28] Concurrently, SAM 3D enables detailed 3D reconstruction of objects and human bodies from single 2D images, including shape, pose, and texture.[52] Self-supervised approaches like DINOv2, introduced in April 2023, produced robust vision encoders from unlabeled data, outperforming supervised models on tasks such as image classification and object detection.[53] DINOv3, scaled in August 2025, further improved performance through larger datasets and refined distillation techniques, achieving state-of-the-art results on benchmarks like ImageNet without task-specific fine-tuning.[54] Reinforcement learning (RL) initiatives target adaptive agents for dynamic environments, particularly in recommendation systems and behavioral modeling. Research integrates RL with graph learning and massive sparse data processing to optimize content ranking on Meta platforms, incorporating techniques like multi-task learning for user behavior prediction.[55] The Pearl library, open-sourced in December 2023, provides tools for off-policy RL evaluation and causal inference, facilitating deployment of RL agents in production settings with verifiable improvements in decision-making efficiency. Efforts in meta-RL explore algorithms that learn adaptation strategies across tasks, as demonstrated in publications advancing RL discovery through automated search, outperforming hand-designed methods in continuous control benchmarks as of October 2025.[56] Embodied AI research focuses on agents interacting with physical and virtual worlds, prioritizing perception and manipulation. The Habitat platform, developed since 2019 and updated through 2025, simulates 3D environments for training navigation and rearrangement agents, enabling zero-shot transfer to real robots via datasets like HM3D. In October 2024, FAIR released open-source advancements in tactile sensing, including the Partnr benchmark for dexterous manipulation and models predicting contact forces from vision and proprioception, aiming to bridge simulation-to-reality gaps in robotics.[57] Meta Motivo, a 2025 behavioral foundation model, generates humanoid actions for virtual agents, supporting multimodal inputs for realistic embodiment in metaverse applications.[58] Multimodal and scientific initiatives extend beyond vision and RL into generative media and domain-specific problem-solving. Movie Gen, introduced in 2025, produces coherent video clips from text prompts using diffusion-based architectures, emphasizing narrative consistency for immersive content creation.[59] In chemistry, the Open Catalyst Project, ongoing since 2020 with expansions in 2025, employs graph neural networks to predict catalyst reactions for sustainable energy, screening millions of candidates to accelerate material discovery over traditional lab methods.[60] Systems research supports these efforts through optimized infrastructure, including custom compilers and distributed computing for scaling multimodal training on Meta's hardware.[61] Despite these outputs, recent restructurings in October 2025 reduced FAIR's headcount by approximately 600 roles, shifting emphasis toward product integration amid competition for resources.[32]Hardware Innovations
MTIA Accelerators and Infrastructure
The Meta Training and Inference Accelerator (MTIA) is a family of custom application-specific integrated circuits (ASICs) developed by Meta to optimize AI workloads, particularly inference for recommendation and ranking models that dominate the company's compute demands. Unlike general-purpose GPUs, MTIA chips are tailored for sparse, high-throughput operations common in Meta's systems, emphasizing cost efficiency and performance for production-scale deployment. The first-generation MTIA (v1), announced on May 18, 2023, marked Meta's entry into custom AI hardware, co-designed alongside PyTorch software and recommendation models to address the limitations of CPU-based servers for growing AI memory and compute needs.[62] MTIA v1 features a architecture optimized for inference, with deployment in Meta's production environments enabling faster processing of ads ranking and content recommendation tasks. This chip integrates into a full-stack solution, reducing reliance on third-party hardware for specific workloads while maintaining compatibility with Meta's software ecosystem. Building on this, the second-generation MTIA (v2), unveiled on April 10, 2024, introduces an 8x8 grid of processing elements delivering 3.5 times the dense compute performance and 7 times the sparse compute performance compared to v1, alongside an upgraded network-on-chip for better scalability. It incorporates 256 MB of on-chip SRAM memory with 2.7 TB/s bandwidth, backed by LPDDR DRAM, prioritizing total cost of ownership (TCO) reductions—up to 44% lower than equivalent GPU setups—through model-chip co-design that aligns hardware directly with Meta's algorithmic needs.[63][64][65] In Meta's infrastructure, MTIA chips form a core component of next-generation data centers, supporting the inference demands of generative AI products, recommendation systems, and ads models across platforms like Facebook and Instagram. As of September 2025, these accelerators are deployed at scale to handle the shift toward AI-driven infrastructure, complementing GPU clusters for training while excelling in real-time inference where sparsity and efficiency yield advantages over commoditized hardware. Meta's approach integrates MTIA into disaggregated compute fabrics, enabling flexible scaling for workloads that process billions of daily predictions. By March 2025, Meta initiated testing of its inaugural in-house training chip, extending the MTIA lineage to full training capabilities and reducing dependency on external suppliers like Nvidia for end-to-end AI pipelines.[66][67][68]Custom AI Hardware Developments
Meta Platforms has expanded its custom AI hardware efforts beyond initial inference-focused accelerators, incorporating training capabilities and strategic acquisitions to optimize large-scale model development. In March 2025, the company began testing its first in-house chip dedicated to AI training, marking a shift from prior emphasis on inference workloads and aiming to enhance efficiency for training expansive models like Llama series.[68] This development, part of the MTIA lineage, targets reduced dependency on third-party GPUs by prioritizing power efficiency tailored to Meta's recommendation and ranking systems.[69] Advancements in model-chip co-design have driven subsequent iterations, with the second-generation MTIA incorporating unified support for PyTorch across hardware types to streamline development. Announced in a June 2025 technical paper, these chips feature enhanced features for handling diverse AI tasks, including doubled performance for recommendation models deployed across platforms like Facebook and Instagram.[70][71] In April 2024, Meta detailed its next-generation MTIA, optimizing software stacks with custom compilers like Triton-MTIA for high-performance code generation on the hardware.[63] To accelerate in-house semiconductor capabilities, Meta acquired Rivos, a startup specializing in AI chip technology, in late September 2025 for an undisclosed sum, integrating its expertise to cut infrastructure costs and lessen reliance on vendors like NVIDIA.[72] This move complements partnerships, such as with Broadcom and Quanta Computer, for deploying next-generation ASIC-powered AI servers announced in August 2025.[73] These efforts reflect a broader infrastructure evolution, blending custom silicon with partner solutions like AMD's MI300 to support escalating AI demands as of September 2025.[66]Products and Deployments
Meta AI Virtual Assistant
Meta AI is a generative AI-powered virtual assistant launched by Meta Platforms in September 2023, initially available in select countries including the United States, and designed to assist users with queries, content creation, and task planning through natural language interactions.[74] It operates as a chatbot integrated directly into Meta's ecosystem, allowing seamless access without requiring a separate download at launch.[6] The assistant is built on Meta's Llama family of large language models, starting with Llama 2 and advancing to Llama 3 in April 2024, Llama 3.1 in July 2024, and Llama 4 in April 2025, which introduced multimodal capabilities for improved voice responses and context retention.[26] [4] The assistant is currently powered by the Llama 4 series, specifically the Scout (17B active parameters, efficient MoE with 16 experts) and Maverick (industry-leading multimodal performance) models, which provide natively multimodal intelligence for text, image understanding, and generation, along with long context support and fast, low-cost inference. Key features of Meta AI include:- Conversational assistance for answering questions, task planning, brainstorming, and personalized recommendations
- Creative tools such as text-to-image generation via "Imagine" and image-to-video animation
- Multimodal capabilities, enabling analysis of images combined with text prompts and generation of visual content
- Unlike Grok, developed by xAI, which is known for its witty, humorous, and maximally truth-seeking style with minimal filtering,
- And ChatGPT from OpenAI, which emphasizes professional, structured, and highly polished responses,
.png)
