AI at Meta

AI at Meta

Research Services

Menlo Park, California 924,934 followers

Together with the AI community, we’re pushing boundaries through open science to create a more connected world.

About us

Through open science and collaboration with the AI community, we are pushing the boundaries of artificial intelligence to create a more connected world. We can’t advance the progress of AI alone, so we actively engage with the AI research and academic communities. Our goal is to advance AI in Infrastructure, Natural Language Processing, Generative AI, Vision, Human-Computer Interaction and many other areas of AI enable the community to build safe and responsible solutions to address some of the world’s greatest challenges.

Industry
Research Services
Company size
10,001+ employees
Headquarters
Menlo Park, California
Specialties
research, engineering, development, software development, artificial intelligence, machine learning, machine intelligence, deep learning, computer vision, engineering, computer vision, speech recognition, and natural language processing

Updates

  • 🎥 Today we’re excited to premiere Meta Movie Gen: the most advanced media foundation models to-date. Developed by AI research teams at Meta, Movie Gen delivers state-of-the-art results across a range of capabilities. We’re excited for the potential of this line of research to usher in entirely new possibilities for casual creators and creative professionals alike. More details and examples of what Movie Gen can do ➡️ https://go.fb.me/00mlgt Movie Gen Research Paper ➡️ https://go.fb.me/zfa8wf 🛠️ Movie Gen models and capabilities • Movie Gen Video: A 30B parameter transformer model that can generate high-quality and high-definition images and videos from a single text prompt. • Movie Gen Audio: A 13B parameter transformer model can take a video input along with optional text prompts for controllability to generate high-fidelity audio synced to the video. It can generate ambient sound, instrumental background music and foley sound — delivering state-of-the-art results in audio quality, video-to-audio alignment and text-to-audio alignment. • Precise video editing: Using a generated or existing video and accompanying text instructions as an input it can perform localized edits such as adding, removing or replacing elements — or global changes like background or style changes. • Personalized videos: Using an image of a person and a text prompt, the model can generate a video with state-of-the-art results on character preservation and natural movement in video. We’re continuing to work closely with creative professionals from across the field to integrate their feedback as we work towards a potential release. We look forward to sharing more on this work and the creative possibilities it will enable in the future.

  • Open Source AI is driving meaningful outcomes across many different platforms and use cases across communities. With their EON models, LinkedIn is able to serve a wide set of platform-specific use cases — while also improving performance in general categories to drive meaningful outcomes for their work. Through their experimentation, they found EON-8B, a domain-adapted version of Llama 3.1 8B, to be 75x and 6x cost effective in comparison to GPT-4 and GPT-4o respectively. More on LinkedIn's domain-adapted foundation models for their platform ➡️ https://go.fb.me/curuy3

    • No alternative text description for this image
  • View organization page for AI at Meta, graphic

    924,934 followers

    New research from Meta FAIR: Large Concept Models (LCM) is a fundamentally different paradigm for language modeling that decouples reasoning from language representation inspired by how humans can plan high-level thoughts to communicate. LCM repo on GitHub ➡️ https://go.fb.me/tkhkeq More details on the Large Concept Models research in the full paper ➡️ https://go.fb.me/0vqbjd

  • Introducing Meta Video Seal: a state-of-the art comprehensive framework for neural video watermarking. Try the demo ➡️ https://go.fb.me/bcadbk Model & code ➡️ https://go.fb.me/7ad398 Details ➡️ https://go.fb.me/n8wff0 Video Seal adds a watermark into videos that is imperceptible to the naked eye and is resilient against common video editing efforts like blurring or cropping, in addition to commonly used compression techniques used when sharing content online. With this release we’re making the Video Seal model available under a permissive license, alongside a research paper, training code and inference code.

  • Open source AI is shaping the future. As we look to 2025, the pace of innovation will only increase as we work to make Llama the industry standard for building on AI. We’ve published a new update on the impact of Llama around the industry and the world ➡️ https://go.fb.me/l1q5nd A few highlights from 2024: 📈 Llama has been downloaded over 650M times. 🌏 License approvals for Llama have more than doubled globally, with significant growth in emerging markets. 🤗 There are now over 85,000 Llama derivative models on Hugging Face alone. ❤️ We’re continuing to see Llama being used across the industry with new examples and innovation from Block, Accenture and LinkedIn — among others. All of this work continues to support our ultimate goal of building the future of human connection and the technology that makes it possible.

  • New research from Meta FAIR — Meta Explore Theory-of-Mind: Program guided adversarial data generation for theory of mind reasoning. This release includes a new research paper, code and a dataset available on Hugging Face. Details on this new work and eight more new releases from FAIR ➡️ https://go.fb.me/ueykrm

  • Fine-tuning Llama 3.1, researchers at MBZUAI (Mohamed bin Zayed University of Artificial Intelligence) built BiMediX2, an Arabic-English VLM for medical use cases that excels in both text and visual tasks — including interpreting medical images from X-Rays, CT scans, MRIs and more. The new model achieves state-of-the-art results in various medical multimodal evaluations thanks to the high-quality bilingual healthcare dataset and instruction sets. More on how they built BiMediX2 with Llama 3.1 ➡️ https://go.fb.me/t9f788 Details in the repo ➡️ https://go.fb.me/ubqbp9 Read the research paper ➡️ https://go.fb.me/crjkch The models and datasets will be available publicly on Hugging Face.

    • No alternative text description for this image

Affiliated pages

Similar pages