12 New Azure Media Services with AI

Microsoft's mission is to empower every person and organization on the planet to achieve more. The media industry is a great example of making this mission a reality. We live in an age where more content is being created and consumed, in more ways and on more devices. At IBC 2019, we shared the latest innovations we're working on and how they can help transform your media experience.
12 New Azure Media Services with AI
Details under the cut!

This page on our website.

Video Indexer now supports animation and multilingual content

Last year at IBC, we made publicly available our award-winning Azure Media Services Video Indexerand this year it got even better. Video Indexer automatically extracts information and metadata such as spoken words, faces, emotions, themes, and brands from media files, and you don't need to be a machine learning expert to use it.

Our latest offerings include previews of two highly requested and differentiated features, Animated Character Recognition and Multilingual Voice Transcription, as well as several additions to the existing models available today in Video Indexer.

Animated Character Recognition

12 New Azure Media Services with AI
Animated content, cartoons are one of the most popular types of content, but standard machine vision models built for human face recognition don't work well with it, especially if the content contains characters without human features. The new preview combines Video Indexer with Microsoft's Azure Custom Vision service to provide a new set of models that automatically detect and group animated characters and make them easy to mark up and recognize with integrated custom vision models.

Models are integrated into a single pipeline, allowing anyone to use this service without any knowledge of machine learning. The results are available through the no-code Video Indexer portal, or through the REST API for quick integration into your own applications.

We created these animated character models with some customers who provided real animated content for training and testing. The value of the new functionality was well described by Andy Gutteridge, senior director of studio technology and post-production at Viacom International Media Networks, who was one of the data providers: β€œAdding a robust AI-based animated content detection feature will allow us to quickly and efficiently find and catalog character metadata from our library. content.

Most importantly, it will give our creative teams the ability to instantly find the content they need, minimize media management time, and allow them to focus on creativity.”

You can get started with the recognition of animated characters with documentation pages.

Identification and transcription of content in multiple languages

Some media resources, such as news, timelines and interviews, contain recordings of people speaking different languages. Most existing speech-to-text translation capabilities require you to specify the audio recognition language first, which makes it difficult to transcribe multilingual videos.

Our new automatic language detection feature for different types of content uses machine learning technology to identify languages ​​found in media resources. Once detected, each language segment automatically goes through a transcription process in the corresponding language, and then all segments are combined into a single transcription file consisting of several languages.

12 New Azure Media Services with AI

The resulting transcript is available as part of the Video Indexer JSON output and as subtitle files. Output transcription is also integrated with Azure Search, allowing you to instantly search your videos for different language segments. In addition, multilingual transcription is available when working with the Video Indexer portal, so you can view the transcript and identified language by time, or jump to specific places in the video for each language and see the multilingual transcription as captions while the video is playing. You can also translate the received text into any of the 54 available languages ​​through the portal and API.

Learn more about the new Multi-Language Content Recognition feature and how it's used in Video Indexer read the documentation.

Additional updated and improved models

We are also adding new models to the Video Indexer and improving existing models, including those described below.

Extracting entities related to people and places

We've expanded our brand detection capabilities to include well-known names and locations such as the Eiffel Tower in Paris and Big Ben in London. When they appear in the generated transcript or on the screen when using Optical Character Recognition (OCR), the appropriate information is added. With this new feature, you can search through all the people, places, and brands that appeared in the video and view details about them, including timeslots, descriptions, and links to the Bing search engine for more information.

12 New Azure Media Services with AI

Frame detection model for editor

This new feature adds a set of "tags" to the metadata attached to individual frames in the details JSON to represent their editorial type (e.g. wide frame, medium frame, close-up, extra close-up, two shots, multiple people, outdoor, indoors, etc.). These frame type characteristics are useful when editing videos for clips and trailers, or when looking for a specific frame style for artistic purposes.

12 New Azure Media Services with AI
Learn more detection of frame types in the Video Indexer.

Advanced IPTC mapping granularity

Our theme detection model determines the theme of a video based on transcription, optical character recognition (OCR) and detected celebrities, even if the theme is not explicitly specified. We map these discovered topics to four classification areas: Wikipedia, Bing, IPTC, and IAB. This enhancement allows us to enable the second level IPTC classification.
Taking advantage of these improvements is as easy as reindexing your current Video Indexer library.

New live streaming functionality

We're also introducing two new features for live streaming in Azure Media Services preview.

Real-time AI transcription takes live streaming to the next level

Using Azure Media Services for live streaming, you can now receive an output stream that includes an auto-generated text track in addition to audio and video content. The text is generated using real-time audio transcription based on artificial intelligence. Custom methods are applied before and after speech-to-text to improve results. The text track is packaged in IMSC1, TTML, or WebVTT, depending on whether it is shipped in DASH, HLS CMAF, or HLS TS.

Real-time line coding for 24/7 OTT channels

Using our v3 APIs, you can create, manage, and live stream over-the-top (OTT) channels, as well as all other features of Azure Media Services, such as live video-on-demand (VOD, video on demand), packaging and digital rights management (DRM, digital rights management).
For previews of these features, visit Azure Media Services communities.

12 New Azure Media Services with AI

New packaging options

Audio track description support

Broadcast content often has an audio track with verbal explanations of what is happening on the screen in addition to the regular audio. This makes programs more accessible to visually impaired viewers, especially if the content is primarily visual. New audio description function allows you to annotate one of the audio tracks as an audio description (AD, audio description) track, so that players can make the AD track available to viewers.

Inserting ID3 metadata

To signal the insertion of ads or custom metadata events to the client's player, broadcasters often use time-based metadata embedded in videos. In addition to the SCTE-35 signaling modes, we now also support ID3v2 and other custom schemes, defined by the application developer for use by the client application.

Microsoft Azure partners demonstrate end-to-end solutions

bitmovin introduces Bitmovin Video Encoding and Bitmovin Video Player for Microsoft Azure. Customers can now use these encoding and playback solutions on Azure and take advantage of advanced features such as 1-step encoding, AVXNUMX/VC codec support, multilingual subtitles, and pre-integrated video analytics for QoS, ads, and video tracking.

Evergent demonstrates its User Lifecycle Management Platform on Azure. As a leading provider of revenue and customer lifecycle management solutions, Evergent uses Azure AI to help premium entertainment providers improve customer acquisition and retention by building targeted service packages and offerings at critical points in their lifecycle.

Havision will showcase its intelligent cloud-based media routing service, SRT Hub, which helps customers transform end-to-end workflows using Azure DataBox Edge and transform workflows with Hublets from Avid, Telestream, Wowza, Cinegy and Make.tv.

SES developed a broadcast-grade media services package on the Azure platform for its satellite and managed media services customers. SES will demonstrate solutions for fully managed playback services, including master playback, localized playback, ad detection and replacement, and 24x7 high-quality real-time multi-channel encoding on Azure.

SyncWords makes convenient cloud-based tools and signature automation technology available on Azure. These offerings will make it easier for media organizations to automatically add captions, including foreign language captions, to live and offline video workflows on Azure.
international company Tata Elxsi, a technology services engineering and delivery company, has integrated its OTT SaaS TEPlay platform with Azure Media Services to deliver OTT content from the cloud. Tata Elxsi also migrated the Falcon Eye QoE monitoring solution to Microsoft Azure, providing analytics and metrics for decision making.

Verizon media makes its live streaming platform available on Azure as a beta. Verizon Media Platform is an enterprise-class managed OTT solution that includes DRM, ad insertion, personalized sessions, dynamic content replacement, and video delivery. Integration simplifies workflows, global support, and scalability, and opens up access to some of the unique capabilities found in Azure.

Source: habr.com

Add a comment