Attention
You have Cookie disabled in your browser. Displaying the site will be incorrect!

Why a streaming project or video hosting service needs computer vision

Automatic moderation of user-generated VOD content
Videos that are prohibited from being published due to ethical, geographic, or other regulations are identified. The analysis allows for the automatic, quick, and accurate identification of the majority of invalid content. A smaller part of the content is sent for manual moderation: The video is tagged with the indication of probabilities and sent for checking by human moderators.
Automatic Live analysis (beta)
Livestreams are constantly analyzed for specified objects. When they appear in a livestream, any restrictive actions can be performed. This allows for the automatic tracking of users’ compliance with content publishing rules.
Content annotation and tagging (beta)
Computer vision (CV) allows you to tag videos based on the identification of scenes, actions, or specified objects. Tags are included in the metadata, and they can serve as the basis for content cataloging or be displayed in video descriptions.
Video markup (beta)
Video can be tagged by the time of the appearance of the specified objects or actions. Thanks to these tags, you can display additional information or enable different types of ads on users’ timelines.

What can be detected with its help

Objects
  • People
  • Faces
  • Pets
  • Household items
  • Logos
  • Vehicles and means of transport
  • Over 1,000 other objects
Объекты
Actions
  • Dancing
  • Eating
  • Fitness
  • Many other actions
Действия
Nudity
  • Female and male faces
  • Covered and exposed body parts
  • Other body parts
Наготу

How to enable and use for VOD and Live

The result of this function is metadata with a list of found objects and the probabilities of their detection.

The result of this function is metadata with a list of found objects and the probabilities of their detection.

Billing plans and expense statistics

Benefits of our solution

5× faster processing speed
Video analysis is done only by key frames and not by the whole video. Video processing time is up to 30× shorter, compared to traditional analysis. The average processing time is 1:5.
Automatic stop triggers
The analysis will stop at the point when the trigger is activated. This allows you to receive answers instantly, without waiting for the complete processing of the video.
Interactive training and functionality improvement
Your project may require customized features. Therefore, the source base of machine learning can be supplemented with your images. We are open to suggestions for integrating new solutions.
Cost optimization
The analysis is faster and does not perform unnecessary actions when detecting the required objects. This saves your budget for the analysis of unnecessary video parts. We also use our own cloud infrastructure with up-to-date technologies.

Compare service features

Gcore AWS Azure Google
Annotation
Objects
Content moderation
VOD processing
Live processing
Cost reduction
Connection of external storage
Quick analysis
Automatic stop triggers

FAQ

How to connect and use computer vision (CV) for VOD and Live?

A full description, use cases, and the necessary documentation can be found in the knowledge base.

Which tools are used for training?

Our models are built on the basis of OpenCV, TensorFlow, and other libraries.

OpenCV is an open-source library for computer vision, image processing, and general-purpose numerical algorithms.

TensorFlow is an open software library for machine learning developed by Google to solve the problems of building and training a neural network in order to automatically find and classify images, achieving the quality of human perception.

What score should I indicate to avoid false positives?

In your videos, CV determines both the objects and the probability of their detection. Each project has its own level of probability, ranging from a slight hint to the impossibility of the appearance of a specified object type. For example, a video has the tag EXPOSED_BREAST_F and a score of 0.51.

How to determine the optimal score?

To determine the average value of your project, we recommend first taking a set of videos (for example, for a day or a week). Then, calculate the score of the specified tags for each video. Lastly, set coefficients based on the result analysis. For example, normal (max. 30%), questionable (max. 50%), and censored (51% and higher).

How to retrain CV that mistakenly misses objects?

We operate with sets of images and videos that cover a large number of uses. However, the system sometimes needs additional training for particular cases.

We recommend generating a set of missed videos and sending them for analysis separately. In the next iteration, the system will also be trained on these videos.

How to process images?

Send images to the system for processing in the same way as videos. A picture is billed as a 1-second video.

How is the cost calculated?

The system takes into account the duration of each processed video in seconds. At the end of the month, the total number is sent to billing. The rate is calculated in minutes.

Let’s say you have uploaded three videos that last 10 seconds, 1 minute and 30 seconds, and 5 minutes and 10 seconds. The sum at the end of the month will be 10 s + 90 s + 310 s = 410 seconds = 6 minutes and 50 seconds. Billing charges 7 minutes. In your personal account, you can see a graph of minutes consumption for each day.

Are videos or images that are used for CV saved?

No. The streaming platform automatically deletes your videos and images after analysis and does not use your data to train basic models. Your video files do not leave your storage and are not sent to edge servers when the container is launched.

Use cutting-edge computer vision and machine learning technologies without investing in infrastructure.

14 days for free • Convenient integration • Efficient 24/7 technical support