The project is based off of a Video-On-Demand architecture builds by AWS which provides the base infrastructure in a CDK-compliant project on to which the streaming service is build.
The API puts a message into a SQS queue which eventually results in the HLS files being put in a destination bucket, triggering a callback to the API containing the same identifier as the API initially provided with the SQS message.
As this project heavily extends on an existing solution, a goal with the implementation for us is to ensure that it is easy to take in upstream changes with minimal conflicts.
We want to be very careful not adding anything too custom, as that adds a potential ton of headache. For the same reason cbx-additions-stack.ts
was added to avoid creating too many conflicts in the existing vod-foundation-stack.ts
.
-
API gateway
DELETE /streaming/{url}
- Deletes S3 stream contentPOST /subtitles
- Converts subtitle files
When a media is deleted any stream and subtitle related to it, is deleted.
The CRON-job
cleanup_eradicated_streams.php
Daily removed all streams, for which the underlying media has been removed.
- Video stream encoding alarm - Reports the AWS-flow failing by monitoring SQS deadletter queue
- Video stream flow callback alarm - Reports the cloud-process being unable to notify the API of a successfully concluded stream job, by monitoring SQS deadletter queue
To find where the encoding-flow might be breaking for a file, the streaming-creation-lifecycle
in prod_logs can give an overview. Similarily for dev videos run through the dev-flow can be tracked in the dev_logs under same-titled table.
The table will always be empty, but intended use it to View in CloudWatch Logs Insights
and here update the unique-media-id in the query, and then scan.
This looks for entries for the media across the multiple logs that span the project. Comparing a "healhty"-job to a "broken"-job makes it much easier to see where it is failing.
The overall subtitle flow is only party owned by this project. The project supports subtitle-conversions through an api-gateway which takes a subtitle and returns a converted, which is used by the API. The storage and use of subtitles is handled else where. More info can be found in the sysadmin wiki subtitles
/source/cdk/cdk.json
- has config for both production (main) and beta (development)job-settings.json
- YAMl-file with encoding configuration. NOTE: THIS IS NOT UPDATED WITH A NEW DEPLOY. YOU HAVE TO MANUALLY SWAP THE FILE FOR A NEW.
Public facing API-gateway, used for subtitle encoding and deleting a stream, is secured by a simple header key x-api-key
The API interacts through both SQS messaging and API-gateways (subtitle management / deletion of streams)
The project is source/destination agnostic in the sense that both the bucket that holds the raw files and the final output bucket are provided as part of the configuration. This made it easy to swap from the previous existing solution without any downtime.
Expected to be plovpenninge
Can be whatever, current setup runs with the original streaming-solution output bucket.
Because Git gets confused
- Branch feature-branch from main
- Write and commit code
- Test deploy to staging through
git push --force origin my-branch:development
- Merge my-branch into master
(Ran from context of the source/cdk folder)
- cdk bootstrap (only needed on first run)
- cdk deploy Note: using this directly is discouraged, use the normal git-procedure and rely on CICD setup instead.
Run the following Athena query to see how many bytes a company have used
SELECT sum(bytes) FROM "cfbillingprod_cf_access_logs_db"."combined" where host = 'd8no9ehofwbfg.cloudfront.net' and (uri LIKE '%AppleHLS1/stream-213697-%' or uri LIKE '%prod_213697-%') and MONTH(date) = 9 and YEAR(date) = 2022
Remember to change 213697
to the correct company ID and the month/year. NOte, if you wonder why we check for two different URIs, it is because an update from AWS (moving it into CDK, etc) changed the format of the URI generated.
streaming-immutable-development
:
Contains items that the CDK deploy doesn't like to attempt to redeploy. As such this is excluded from the pipeline but may become an unnecessary part in the future.
streaming-defaults-development
:
The original AWS CDK project that was forked and adapted. Try to keep changes to this at a minimum to ease upstream merges. This basically contains the entire flow of moving a file from a folder, converting and moving result to a final cloudfront-backed destination folder.
streaming-custom-development
:
Contains all the custom additions used to communicate between the API and the default-solution a long with subtitles, cleanup and more.
Just like any cloudformation project, removing the project will remove all components. Storage is usually persisted and "orphaned" after delete, unless the component has its lifecycle specifically configured to die (and delete content) with the stack.
Application has from AWS-authors 2 custom-resource-backed lambdas. These currently (2021-08-26) don't include support for their own teardown, so their deletion processes will hang (time out after 1 hour) and the deletion will fail. To fix this, follow this: https://www.youtube.com/watch?v=hlJkMoCxR-I
For cases where existing development setup isn't sufficient, it's also an option to spin up a new stack.
-c destination_bucket_name=claus-destination
-c api_host=claus-api.cbx.xyz
-c stream_host=https://dpv7mez7ofol6.cloudfront.net
cdk deploy -c destination_bucket_name=claus-destination -c api_host=claus-api.cbx.xyz stream_host=https://dpv7mez7ofol6.cloudfront.net --all
cdk synth -c destination_bucket_name=claus-destination -c api_host=claus-api.cbx.xyz stream_host=https://dpv7mez7ofol6.cloudfront.net --all
See source/custom-resource/lib/s3/index.js
Resource buckets very likely a plovpenninge bucket
Bucket-permissions -> Bucket-policy:
{
"Version": "2008-10-17",
"Id": "PolicyForCloudFrontPrivateContent",
"Statement": [
{
"Sid": "1",
"Effect": "Allow",
"Principal": {
"AWS": "arn:aws:iam::cloudfront:user/CloudFront Origin Access Identity <IDENTITY>"
},
"Action": [
"s3:GetObject*",
"s3:GetBucket*",
"s3:List*"
],
"Resource": [
"arn:aws:s3:::<bucket-name>",
"arn:aws:s3:::<bucket-name>/*"
]
}
]
}
Bucket-permissions -> Bucket-CORS:
[
{
"AllowedHeaders": [
"*"
],
"AllowedMethods": [
"GET"
],
"AllowedOrigins": [
"*"
],
"ExposeHeaders": [],
"MaxAgeSeconds": 3000
}
]
How to implement a video-on-demand workflow on AWS leveraging AWS Lambda, AWS Elemental MediaConvert, Amazon s3 and Amazon CloudWatch. Source code for Video on Demand on AWS Foundation solution.
The AWS CloudFormation template deploys a workflow that ingests source videos, transcodes the videos into multiple Adaptive Bitrate Formats (ABR) and delivers the content through Amazon CloudFront. The solution creates a source Amazon S3 bucket to store the source video files, and a destination bucket to store the outputs from AWS Elemental MediaConvert. A job-settings.json file, used to define the encoding settings for MediaConvert, is uploaded to the source S3 bucket.
The solution includes two AWS lambda functions: a job submit function to create the encoding jobs in MediaConvert and a job complete function to process the outputs. Amazon CloudWatch tracks encoding jobs in MediaConvert and triggers the Lambda job complete function. An Amazon SNS topic is deployed to send notifications of completed jobs, and Amazon CloudFront is configured with the destination S3 bucket as the origin for global distribution of the transcoded video content.
For more detail including using your own settings file please see the solution implementation guide
The solution can be deployed through the CloudFormation template available on the solution home page: Video on Demand on AWS.
The solution was developed using the AWS Cloud Development Kit (CDK) and leverage 3 of the AWS Solutions Constructs. To make changes to the solution, download or clone this repo, update the source code and then either deploy the solution using the CDK or run the deployment/build-s3-dist.sh script. The build script will generate the CloudFormation template from the CDK source code using cdk synth, run deployment/cdk-solution-helper to update the template to pull the lambda source code from S3 and package the Lambda code ready to be deployed to an Amazon S3 bucket in your account.
For details on deploying the solution using the CDK see the CDK Getting Started guide
- AWS Command Line Interface
- Node.js 12.x or later
- aws-cdk version 1.63.0
Run unit tests to make sure added customization passes the tests:
cd source/custom-resource
npm install
cd ../job-submit
npm install
# Then go back to deployment directory
cd ../../deployment
chmod +x ./run-unit-tests.sh
./run-unit-tests.sh
The CloudFormation template is configured to pull the Lambda deployment packages from Amazon S3 bucket in the region the template is being launched in. Create a bucket in the desired region with the region name appended to the name of the bucket (e.g. for us-east-1 create a bucket named my-bucket-us-east-1
).
aws s3 mb s3://my-bucket-us-east-1
Build the distributable:
chmod +x ./build-s3-dist.sh
./build-s3-dist.sh my-bucket video-on-demand-on-aws-foundation v1.0.0
Notes: The build-s3-dist script expects the bucket name as one of its parameters, and this value should not include the region suffix.
Deploy the distributable to the Amazon S3 bucket in your account:
aws s3 cp ./regional-s3-assets/ s3://my-bucket-us-east-1/video-on-demand-on-aws-foundation/v1.0.0/ --recursive --acl bucket-owner-full-control
- Deploy the cloudFormation template from deployment/global-assets/video-on-demand-on-aws-foundation.template into the same region as you newly created S3 bucket.
Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.
This solution collects anonymous operational metrics to help AWS improve the quality of features of the solution. For more information, including how to disable this capability, please see the implementation guide.