Postitatud - Lisa kommentaar

EgoMimic: Georgia Tech doktorant kasutab Project Aria teadusprille, et aidata treenida humanoidseid roboteid

Today, we’re highlighting new research from Georgia Tech that helps train robots to perform basic everyday tasks using egocentric recordings from wearers of Meta’s Projekt Aria research glasses. Check out the video below, read the full story, or apply for your own Project Aria Research Kit.

Imagine having help completing everyday tasks in your home such as doing the laundry, washing dishes, and making repairs. We already use tools to help with these tasks, like washing machines, dishwashers, and electric drills. But what if you could have an even more powerful and flexible tool in the form of a humanoid robot that could learn from you and accelerate any number of physical projects on your to-do list?

Even if you had the available hardware system, teaching a robot to do everyday tasks can only be achieved through a slow and clunky data collection method called robot teleoperation. Until now. By using the Project Aria Research Kit, Professor Danfei Xu and the Robotic Learning and Reasoning Lab at Georgia Tech use the egocentric sensors on Aria glasses to create what they call “human data” for tasks that they want a humanoid robot to replicate. They use human data to dramatically reduce the amount of robot teleoperation data needed to train a robot’s policy—a breakthrough that could some day make humanoid robots capable of learning any number of tasks a human could demonstrate.

Kareer teleoperates the robot to capture co-training data for EgoMimic. Teleoperation can be difficult to scale and require significant human effort.

“Traditionally, collecting data for robotics means creating demonstration data,” says Simar Kareer, a PhD student in Georgia Tech’s School of Interactive Computing. “You operate the robot’s joints with a controller to move it and achieve the task you want, and you do this hundreds of times while recording sensor data, then train your models. This is slow and difficult. The only way to break that cycle is to detach the data collection from the robot itself.”

Today, robot policy models are trained with large amounts of targeted demonstration data specific to each narrow task at a high cost. Kareer hypothesizes that passively collected data from many researchers, like the data captured by Aria glasses, could instead be used to enable data creation for a much broader set of tasks to create more generally useful robots in the future.

Inspired by Projekt Aria ja Ego-Exo4D which includes a massive egocentric dataset of over 3K hours of video recordings of daily-life activities, Kareer developed EgoMimic, a new algorithmic framework that utilizes human data and robot data for humanoid robot development.

“When I looked at Ego4D, I saw a dataset that’s the same as all the large robot datasets we’re trying to collect, except it’s with humans,” Kareer explains. “You just wear a pair of glasses, and you go do things. It doesn’t need to come from the robot. It should come from something more scalable and passively generated, which is us.” In Kareer’s research, Aria glasses were used to create human data for co-training the EgoMimic framework.

Kareer creates co-training human data by recording with Aria glasses while folding a t-shirt.

Aria glasses aren’t just used for human data collection in Georgia Tech’s research. They’re also used as an integral component of the robot’s real-time operation setup. Aria glasses are mounted to their humanoid robot platform just like a pair of eyes and serve as an integrated sensor package that enables the robot to perceive its environment in real time. The Aria Client SDK is utilized to stream Aria’s sensor data directly into the robot’s policy, running on an attached PC, which in turn controls the robot’s actuation. Using Aria glasses for both the data collection and the real-time perception pipeline minimizes the domain gap between the human demonstrator and the robot, paving the way for scaled human data generation for future robotics task training.

Aria glasses mounted to the top of the robot provide the system with sensor data that allows the robot to perceive and interact with the space.

Thanks to EgoMimic, Kareer achieved a 400% increase in his robot’s performance across various tasks vs previous methods with just 90 minutes of Aria recordings. The robot was also able to successfully perform these tasks in previously unseen environments.

In the future, humanoid robots could be trained at scale using egocentric data in order to perform a variety of tasks in the same way humans do.

“We look at Aria as an investment in the research community,” says James Fort, a Reality Labs Research Product Manager at Meta. “The more that the egocentric research community standardizes, the more researchers will be able to collaborate. It’s really through scaling with the community like this that we can start to solve bigger problems around how things are going to work in the future.”

Kareer will present his paper on EgoMimic at the 2025 IEEE Engineers’ International Conference on Robotics and Automation (ICRA) in Atlanta.

Postitatud - Lisa kommentaar

Aria Gen 2 tutvustamine: uute teadusuuringute avamine masintaju, kontekstuaalse AI, robootika ja muu vallas

Alates selle käivitamisest 2020, Projekt Aria on võimaldanud teadlastel kogu maailmas arendada masinataju ja tehisintellekti tipptasemel teadusriistvara ja avatud lähtekoodiga tarkvara kasutamise kaudu. andmekogumid, mudelid ja tööriistad. Täna on meil hea meel teatada selle teekonna järgmisest sammust: Aria Gen 2 prillide tutvustamine. See järgmise põlvkonna riistvara avab uued võimalused paljudes uurimisvaldkondades, sealhulgas masintaju, egotsentriline ja kontekstuaalne tehisintellekt ja robootika.

0:00 / 0:00

Teadlastele, kes soovivad uurida, kuidas tehisintellekti süsteemid saavad paremini mõista maailma inimese vaatenurgast, Aria Gen 2 prillid lisavad Aria platvormile uusi võimalusi. Need sisaldavad mitmeid uuendusi, mida ei leidu ühelgi teisel praegu saadaval oleval seadmel, ning juurdepääs nendele läbimurdelistele tehnoloogiatele võimaldab teadlastel avardada võimaluste piire.

Võrreldes Aria Gen 1-ga sisaldab Aria Gen 2 ainulaadne väärtuspakkumine järgmist:

  • Moodne andurikomplekt: Uuendatud sensorikomplektis on RGB-kaamera, 6DOF SLAM-kaamerad, silmade jälgimise kaamerad, ruumilised mikrofonid, IMU-d, baromeeter, magnetomeeter ja GNSS. Võrreldes oma eelkäijaga, Aria Gen 1, on uues põlvkonnas kasutusele võetud kaks uuenduslikku ninaosa sisseehitatud andurit: PPG-sensor südame löögisageduse mõõtmiseks ja kontaktmikrofon, mis eristab kandja häält kõrvalseisjate häälest.
  • Ülimalt madala energiatarbega ja seadmesisene masinataju: SLAM, silmade jälgimine, käe jälgimine ja kõnetuvastus töödeldakse seadmes, kasutades Meta kohandatud räni.
  • Kogu päeva kasutatavus: Aria Gen 2 prillid on võimelised kuus kuni kaheksa tundi pidevat kasutamist, kaaluvad umbes 75 grammi ja neil on kokkupandavad käed, et neid oleks lihtne transportida.
  • Interaktsioon heli kaudu: Kasutajad saavad helitagasisidet oma klassi parimate avatud kõrvakõlarite kaudu, mis võimaldab kasutaja-ühenduses süsteemi prototüüpimist.

Meie kümnendi pikkune teekond järgmise arvutiplatvormi loomiseks on viinud nende kriitiliste tehnoloogiate väljatöötamiseni. Meta, Reality Labs Researchi ja FAIR AI labori meeskonnad kasutavad neid, et edendada meie pikaajalist teadusvisiooni.. Nende kättesaadavaks tegemine akadeemilistele ja kommertslaboratooriumidele projekti Aria kaudu edendab veelgi avatud teadusuuringuid ja üldsuse arusaamist võtmetehnoloogiatest, mis meie arvates aitavad kujundada arvutite ja tehisintellekti tulevikku.

Projekti Aria poolt alates 2020. aastast võimaldatud avatud teadusuuringud on juba viinud olulise töö tegemiseni, sealhulgas avatud lähtekoodiga vahendite loomiseni, mida kasutatakse laialdaselt nii akadeemilistes ringkondades kui ka tööstuses. . Ego-Exo4D andmekogum, mis on kogutud esimese põlvkonna Aria prillidega, on muutunud põhiliseks töövahendiks kaasaegses arvutinägemises ja kasvavas robootika valdkonnas. Georgia Techi teadlased näitasid hiljuti, kuidas Aria uurimiskomplektiga saab aidata humanoidrobotitel õppida inimesi kodus abistama, samas kui BMW meeskonnad kasutas seda selleks, et uurida, kuidas integreerida liitreaalsuse ja virtuaalse reaalsuse süsteeme arukatesse sõidukitesse.

Aria võimaldab ka uute tehnoloogiate arendamist ligipääsetavuse tagamiseks. Esimese põlvkonna Aria prille kasutas Carnegie Melloni Ülikooli NavCogi projektis, mille eesmärk oli luua tehnoloogiaid, mis aitaksid pimedatel ja vaegnägijatel siseruumides navigeerida. Sellele alusele tuginedes kasutavad Aria Gen 2 prille nüüd järgmised ettevõtted. Envisionettevõte, mis on pühendunud pimedatele või vaegnägijatele mõeldud lahenduste loomisele. Envision uurib oma toodete integreerimist Ally Tehisintellekti assistent ja ruumiline heli, mis kasutab uusimaid Aria Gen 2 prille, et parandada siseruumides navigeerimist ja ligipääsetavust.

0:00 / 0:00

Envision kasutas Aria Gen 2 seadme SLAM-funktsioone koos ruumilise audiofunktsiooniga, mis on sisseehitatud kõlarite kaudu, et aidata pimedatel ja vaegnägijatel sujuvalt liikuda siseruumides. See innovatiivne tehnoloogiate kasutamine, mis on veel uurimis- ja uurimisfaasis, näitab, kuidas teadlased saavad kasutada Aria Gen 2 prille egotsentrilistel vaatlustel põhinevate tehisintellekti kogemuste prototüüpimiseks. Täiustatud sensorid ja seadme masinataju võimalused, sealhulgas SLAM, silmade jälgimine, käe jälgimine ja audiointeraktsioonid, muudavad need ideaalseks ka andmete kogumiseks teadusuuringute ja robootikarakenduste jaoks.

Järgnevate kuude jooksul jagame partneritele rohkem üksikasju seadme kättesaadavuse ajastuse kohta. Aria Gen 2 kasutamisest huvitatud teadlased saavad kasutada järgmist registreeru, et saada uuendusi. Oleme põnevil, et näha, kuidas teadlased kasutavad Aria Gen 2, et sillutada teed tulevastele uuendustele, mis kujundavad järgmise arvutiplatvormi.

Postitatud - Lisa kommentaar

Inside Aria Gen 2: Explore the Cutting-Edge Tech Behind the Device

Earlier this year, we announced our latest research glasses, Aria Gen 2, marking the continuation of Project Aria’s mission to enable researchers across the world to advance the state of the art in machine perception, contextual AI, and robotics through access to cutting-edge research hardware and open source datasets, models, and tooling. Today, we’re excited to share more about the technology inside Aria Gen 2. This includes an in-depth overview of the form factor, audio capabilities, battery life, upgraded cameras and sensors, on-device compute, and more.

What Is Aria Gen 2?

Aria Gen 2 is a wearable device that combines the latest advancements in computer vision, machine learning, and sensor technology. Aria Gen 2’s compact form factor and lightweight design make it an ideal choice for researchers who need to collect data or build prototypes in a variety of settings. The glasses contain a number of improvements when compared to Aria Gen 1, its research predecessor, announced back in 2020.

Aria Gen 2: Advancements and Features

The transition from Aria Gen 1 to Gen 2 marks a significant leap in wearable technology, offering enhanced features and capabilities that cater to a broader range of applications and user needs. Below, we explore the key differences and improvements introduced in Aria Gen 2.

1. Wearability

Aria Gen 2 boasts superior wearability, characterized by enhanced comfort and fit, while accommodating a wider range of face morphologies and a rich sensor suite for research. The glasses maintain a lightweight design (weighing in at 74 – 76g, depending on size) and now include folding arms for easier storage and transport for everyday use. To ensure each wearer has an optimal physical and functional fit, we’ve introduced eight size variations of the device—accounting for a number of human factors including head breadth and nose bridge variation.

Eight size variations of our Aria Gen 2 devices.

2. Computer Vision (CV) Camera Enhancements

High Dynamic Range (HDR): Aria Gen 2’s global shutter camera sensor offers a high dynamic range of 120 dB, compared to the 70 dB range in Gen 1. This allows for superior computer vision tasks across diverse lighting conditions.

0:00 / 0:00

The video illustrates that the CV camera is able to capture highly dynamic scenes with an LED light’s filament being resolved along with the rest of the details from the scene.

Wide Field of View (FOV): Aria Gen 2 is equipped with four computer vision (CV) cameras, doubling the number of CV cameras in Gen 1, to provide a wider field of view and enable advanced 3D hand and object tracking.

Stereo Overlap: The stereo overlap in Gen 2 is increased to 80° from Gen 1’s 35°, facilitating stereo-based foundation models that enhance depth perception and spatial awareness.

0:00 / 0:00

The example here illustrates how the increased stereo overlap enables methods such as NVIDIA’s FoundationStereo to generate depth maps based on rectified stereo images. The depth maps can be fused to generate geometric reconstructions of the scene only using Aria Gen 2’s stereo pair data.

3. New Sensor Integrations

Ambient Light Sensor (ALS): Aria Gen 2 includes a calibrated ALS, enabling better exposure control algorithms and unlocking new capabilities at low frame rates. The ALS’s ultraviolet mode can be used to distinguish between indoor and outdoor lighting as illustrated by the video.

0:00 / 0:00

Contact Microphone: Aria Gen 2 includes a contact microphone embedded in the nosepad of the device, enhancing audio capture in noisy environments.

0:00 / 0:00

The video showcases a wearer in a wind tunnel to simulate a windy scenario where the contact microphone is able to pick up the wearer’s whisper when the acoustic microphones cannot.

Heart Rate: Aria Gen 2 includes a photoplethysmography (PPG) sensor embedded in the nosepad of the device, that enables estimation of heart rate of the person wearing the device.

4. Device Time Alignment

Aria Gen 2 has an onboard hardware solution that utilizes Sub-GHz radio technology to broadcast timing information, enabling precise time alignment with other Aria Gen 2 devices or compatible devices that support Sub-GHz radio. This technology achieves time alignment with an accuracy of sub-millisecond, marking a significant improvement over the software-based alignment of Gen 1.

0:00 / 0:00

The video shows how Aria Gen 2 uses device time alignment for tasks like writing from distributed captures from two Aria Gen 2 devices.

5. On-device Realtime Machine Perception (MP) Signals

Aria Gen 2 features advanced on-device machine perception algorithms that run on Meta’s energy-efficient custom coprocessor. These cutting-edge capabilities enable the device to generate precise and accurate data, tracking how we interact with our surroundings.

Visual Inertial Odometry (VIO)

One of the key features of Aria Gen 2 is its ability to track the glasses in six degrees of freedom (6DOF) within a spatial frame of reference using Visual Inertial Odometry (VIO). This allows for seamless navigation and mapping of the environment, opening up new possibilities for research in contextual AI and robotics.

Eye Tracking

Aria Gen 2 also boasts an advanced camera-based eye tracking system that tracks the wearer’s gaze with unparalleled accuracy. This system provides a wealth of information, including: gaze per eye, vergence point, blink detection, pupil center estimation, pupil diameter, corneal center, etc.

These advanced signals enable a deeper understanding of the wearer’s visual attention and intentions, unlocking new possibilities for human-computer interaction.

Hand Tracking

Aria Gen 2 also features a hand tracking solution that tracks the wearer’s hand in 3D space. This produces articulated hand-joint poses in the device frame of reference, facilitating accurate hand annotations for datasets and enabling applications such as dexterous robot hand manipulation that require high precision.

0:00 / 0:00

Demonstration of Aria Gen 2’s sensors and machine perception capabilities, as well as off-device algorithms built on them.

The Future of Aria Is Here: Stay Informed

Aria Gen 2 glasses pave the way for future innovations that will define the next computing platform. Applications to work with Aria Gen 2 will open later this year, and researchers who are interested in staying informed can join the Aria Gen 2 interest list. Meanwhile, applications for Aria Research Kit with Aria Gen 1 glasses are still being accepted on a rolling basis—apply now to get started immediately.

Join us at CVPR 2025 in Nashville, Tennessee, this June, where the team will showcase Aria Gen 2 glasses through interactive demos. Visit the Meta booth to experience the latest advancements and learn more about the innovative features of Aria Gen 2.