[Beowulf] transcode Similar Video Processing on Beowulf?

Mark Hahn hahn at mcmaster.ca
Wed Apr 16 09:54:55 PDT 2014

I'm trying to understand this from a perspective of conventional HPC.

> cop-out but we're not keen to reinvent the wheel. It provides
> statekeeping and job queues in one package; replacing it wouldn't be

"statekeeping" is just tracking queued/running/done jobs, right?

> trivial but wouldn't be a massive task; the cost of using it is tiny,
> though, and it made our life a lot easier. It's all written in terms
> of deciders, which make decisions based on a list of events associated
> with an event (eg a "finished activity" event will have the details
> about the activity starting, being scheduled, and being completed,
> output status etc),

is the workflow complicated - a directed graph with complicated 
structure, rather than a series of discrete jobs, each a simple 
chain/pipeline in structure?

> maintained by passing JSON blobs around as messages; there'll be a
> blog post or two explaining things on our website soonish and I'll
> post them across if there's interest.

a reference would be interesting.

> It's being used in production on a regular basis and has had quite a
> lot of content processed through it so far; these tasks on average run
> for 2-6 hours and involve ~1GB of data going in and a few megabytes
> out.

that's unexceptional from an HPC perspective.

> The APIs are all simple HTTPS RESTful ones, storage can be cloud
> provider storage or local shared drive storage.

one premise usually found in HPC is that the job, at least the main part,
should be compute-bound.  how do you ensure that your compute resources
are not idle or starved by external IO bottlenecks?

> interprocess communication performance is less important and
> robustness and dynamic scalability plays a major role.

well, I think that's a bit disingenuous, since HPC is highly tuned
for robustness and dynamic scalability...

thanks, mark hahn.

More information about the Beowulf mailing list