Xinhua, China’s state-run press agency, has unveiled new “AI anchors” — digital composites created from footage of human hosts that read the news using synthesized voices.
It’s not clear exactly what technology has been used to create the anchors, but they’re in line with the most recent machine learning research. It seems that Xinhua has used footage of human anchors as a base layer, and then animated parts of the mouth and face to turn the speaker into a virtual puppet. By combining this with a synthesized voice, Xinhua can program the digital anchors to read the news, far quicker than using traditional CGI. (We’ve reached out to AI experts in the field to see what their analysis is.)
According to reports from Xinhua and the South China Morning Post, two anchors (one for English broadcasts and one for Chinese) were created in collaboration with local search engine company Sogou. Xinhua says the anchors have “endless prospects” and can be used to cheaply generate news reports for the agency’s TV, web, and mobile output.
Each anchor can “work 24 hours a day on its official website and various social media platforms, reducing news production costs and improving efficiency,” says Xinhua.
The technology has its limitations. In the videos above and below of the English-speaking anchor, it’s obvious that the range of facial expressions are limited, and the voice is clearly artificial. But machine learning research in this area is making swift improvements, and it’s not hard to imagine a future where AI anchors are indistinguishable from the real thing.
Click here to read more.
SOURCE: The Verge, James Vincent