Native AI Models Learn From Scratch With Minimal Data Required

Scientists created a new family of vision-language models from scratch using simpler building blocks, allowing them to develop visual perception without the need for vast amounts of training data. This breakthrough has the potential to accelerate progress in the field by making it easier and more cost-effective for researchers to build powerful models that can understand both images and text.

Read Paper →

AI Learns to Build Machines with Ease

Scientists built a testbed using a game to see if large language models can learn to create complex machines by assembling parts. If successful, this could lead to more advanced artificial intelligence that can understand and interact with the physical world in a more human-like way.

Read Paper →
a person's hand on a wall with a shadow cast on it

AI Builds Accurate 3D Environments with Native Data

We created a new way to build 3D models of the world that don't rely on flat images, but instead use three-dimensional data to create more accurate and efficient representations. This could lead to better computer-generated environments for things like virtual reality or video games, where it's important to be able to see objects from any angle without them looking distorted or fake.

Read Paper →

AI Learns to See People from Multiple Angles

We created a large dataset of images with multiple views of the same person to help artificial image generators produce more varied and realistic results,.

Read Paper →