To develop its lip-reading AI, Meta uses AV-HuBERT, a unique multimodal learning system that combines audio and lip-movement cues to perceive language. ![]() But let's see what Meta has been working on lately. ![]() ![]() She also pointed out that Meta has introduced four responsible innovation principles for development "with ethics, privacy, safety, and security at the forefront." Coming from a company that in the past has shown very little regard for the privacy and ethical concerns of its users, however, we can't say we are convinced. Kristen Morea, a Meta spokeswoman, reported that the company has made $50 million worth of investments in external programs so far to ensure that the Metaverse is built safely. But, thanks to Meta's Audio-Visual Hidden Unit BERT (AV-HuBERT) framework, which is learning to understand language by both listening and watching people communicate, this is about to change. Until now, even the most advanced AI was only successful in recognizing verbal cues (in contrast to humans which use visual ones, like lip movement, facial expressions, and hand gestures, as key elements in language learning). Effective communication involves speech, gestures, tone, and more – essentially, verbal and non-verbal elements.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |