File: //usr/lib/python3/dist-packages/awscli/examples/rekognition/start-content-moderation.rst
**To start the recognition of unsafe content in a stored video**
The following ``start-content-moderation`` command starts a job to detect unsafe content in the specified video file stored in an Amazon S3 bucket. ::
aws rekognition start-content-moderation \
--video "S3Object={Bucket=MyVideoS3Bucket,Name=MyVideoFile.mpg}"
Output::
{
"JobId": "1234567890abcdef1234567890abcdef1234567890abcdef1234567890abcdef"
}
For more information, see `Detecting Unsafe Stored Videos <https://docs.aws.amazon.com/rekognition/latest/dg/procedure-moderate-videos.html>`__ in the *Amazon Rekognition Developer Guide*.