Multimedia Editing in the Cloud: Treating Audio as Big Data Adam - - PowerPoint PPT Presentation

multimedia editing in the cloud treating audio as big data
SMART_READER_LITE
LIVE PREVIEW

Multimedia Editing in the Cloud: Treating Audio as Big Data Adam - - PowerPoint PPT Presentation

Multimedia Editing in the Cloud: Treating Audio as Big Data Adam T. Lindsay Multi-Service Networks, Coseners House, 2009 and now for something completely different... Some context Im interested in Metadata-assisted multimedia


slide-1
SLIDE 1

Multimedia Editing in the Cloud: Treating Audio as ‘Big Data’

Adam T. Lindsay Multi-Service Networks, Cosener’s House, 2009

slide-2
SLIDE 2

and now for something completely different...

slide-3
SLIDE 3

I’m interested in… Metadata-assisted multimedia editing Using music analysis service to get sample- accurate, hierarchical event pointers Using these for music remixing

Some context

slide-4
SLIDE 4

Remixing using metadata

Tatums Beats Bars Song

slide-5
SLIDE 5

flickr.com/photos/meganpru/455156509/

slide-6
SLIDE 6

Make the programming paradigm as declarative/functional as possible Rendering output can be done independently

  • f handling metadata

A content description can act as proxy for the underlying content Don't need to handle the data at all Rendering instructions form a small vocabulary

Splitting things into tiny pieces

slide-7
SLIDE 7

Smells a lot like SMIL 1.0

Rendering

<sequence duration="57.11676" source="847e7a3146fb790ccfa4a071f7395775"> <trackinfo filename="../music/aha.mp3" id="847e7a3146fb790ccfa4a071f7395775"/> <trackinfo filename="../music/SLadies.mp3" id="1630307ae0eea4a380ab2213827eec6f"/> <trackinfo filename="../music/BJean.mp3" id="2d539b439ec027e73abd2390c5611d2f"/> <parallel duration="0.33472"> <beat duration="0.33472" start="0.21285"/> <beat duration="0.31216" source="1630307ae0eea4a380ab2213827eec6f" start="0.38352"/> </parallel> <parallel duration="0.52013" source="2d539b439ec027e73abd2390c5611d2f"> <beat duration="0.52013" start="0.70155"/> <beat duration="0.3355" source="847e7a3146fb790ccfa4a071f7395775" start="0.54757"/> </parallel> </sequence>

slide-8
SLIDE 8

Transparent Linear superposition of waveforms = mixing Simple information set: Source, source-start-time, source-duration Destination, destination-start-time Return samples, destination-start-time

Audio is easy

slide-9
SLIDE 9

Feels like MapReduce Data-intensive (CD audio = 10 MB/min) Need a strategy to cope with data-heavy nature One Map task per source One Reduce per job Collection/Reduce best at biggest contributing source

Source independence

slide-10
SLIDE 10

Proof of concept Using small-scale (flat) P2P network (Based on rift libraries for RPC implementation) Consistent hashing for matching content with node No experiments to test gains and costs yet

Implementation

slide-11
SLIDE 11
slide-12
SLIDE 12

Video can fit into this model too ...if we apply alpha channel adjustments first Editing in the local network Push content addressing down network stack Maybe push aggregation to content processor nodes

e Future

slide-13
SLIDE 13

Frame/sample-accurate fragment addressing Better content-centric addressing Some access control/billing models

Mixing on a web scale

slide-14
SLIDE 14

Fin