# gc3pie

34 thoughts
last posted Jan. 3, 2013, 4:26 p.m.
0
get stream as: markdown or atom
0

Just produced a Littlewood fractal of (18 degrees/400 size) but it seems skewed, perhaps I made an error in transposing the code when moving to scripts:

0

I wrote on an earlier card how I needed to add:

if [[ $- != *i* ]]; then . /etc/profile.d/sge.sh fi  To my ~/.bashrc but failed to mention it needs to be on the sgeadmin user's account (or whatever user you have configured to connect with). 0 Whatever local files you pass to the inputs parameter of the Application instance, will get pushed to the remote server, this includes any binary or script that you are going to execute. It's not just limited to data files used for input. 0 I think the goal (need clarification) of using models/Django is record what was run, when, and with what inputs (as well as what outputs, final and intermediary) were generated. 0 The trick is how we marry this construction of workflows up with Django models. I think it might be a little far reaching to make it complete data driven. 0 For every distinct binary (or binary with certain parameters), I think an Application subclass will be created in a library of apps that can be used to construct Workflows of various types of collections that glue these apps together funneling the output from some into the input of others and making decisions about what next to run in the SequentialTaskCollection tasks. 0 From part11.pdf: After a task has completed, the next method is called with the index of the finished task in the ‘self.tasks‘ list the return value of the next method is then made the collection execution.state. If the returned state is ‘RUNNING‘, then the subsequent task is started, otherwise no action is performed. 0 Each collection specifies a task list in the __init__ constructor. You can also implement a next() method that will be called after each task completes. You have access to this list and can dynamically pick the next Application to run from within this method. 0 You compose workflows by combining the various Applications (or executable bits that will run on a node in the cluster) to either run in parallel or serial, and these collections can run in parallel or serial all the way up until you have a single collection that will run everything. This top most collection is the Workflow. 0 The Workflow isn't really it's own object but the top most Collection that owns all other collections. 0 The *Collection objects can other *Collections as well as Applications. 0 The main objects to work with in gc3pie are: • Task • Application • SequentialCollection • ParallelCollection • Workflow • SessionBasedScript 0 The coolest part, I think, is how you can dynamically influence the next steps, either by changing parameters based on previous output, or selecting different jobs entirely at runtime. It's all about how you write your next() method on your workflow objects. For example: 0 So after running apt-get install -y imagemagick across all nodes in my cluster, the warholize script ran against: to produce: By running: python warholize.py -C 1 ~/Pictures/paltman.png --copies 9  Using all 9 nodes of my cluster in parallel following the following workflow: 0 I think the warholize demo might provide deeper insight into how all this works together with a bit more real-world example. 0 Have successfully working through most of the 2 exercises at the end of part03's slide deck, however, having a bit of trouble now in understanding how the postprocess/terminate steps work and how their output makes their way back to the consumer. 0 There was a minor change I needed to make to squares.py as it was causing a syntax error on the node of execution: On line 33, I changed:  arguments=['expr', x, '*', x],  to:  arguments=['expr', x, '\*', x],  In order to escape the * character so that it would be interpreted properly by the shell. 0 So the environmental issues were caused by the StarCluster image loading SGE environment variables on login shells instead of through both login and non-login shells. I added this to my user's ~/.bashrc script and all is working now: if [[$- != *i* ]]; then
. /etc/profile.d/sge.sh
fi

0

So close.

Dealing with environmental issues, now.

It connects fine, however, when I connect manually via SSH with the same user/key, I have a complete environment (all the expected env variables are set). When the script connects with SshTransport it's throwing exceptions and error messages that indicate that the environment variables are not set.

0

Now that this is available (starcluster start mycluster is pretty awesome for on demand compute clusters), I am going to work through the exercises at the end of part03.

0

Firing up an SGE cluster this morning on EC2 using StarCluster to have to work through some of the tutorials and exercises in GC3Pie.

0

This slide deck explains things pretty well.

starting to click...

0

Same for listing the outputs. This list (passed to the Application constructor) will be the files that will be copied back locally once the execution completes.

0

For instance, with SAGA-python you have to explicitly copy over your input files before starting to execute anything. With GC3pie, you just list your local input files along with the executable that you want to execute remotely, and it handles copying the files to the remote server where the execution will happen.

0

All the examples in the tutorials are calling the base class methods directly rather than using super(), for example:

class SquareApplication(Application):
def __init__(self, x):
self.to_square = x
Application.__init__(self, ...)


class SquareApplication(Application):
def __init__(self, x):
self.to_square = x
super(SquareApplication, self).__init__(...)


I wonder why this is done.

0

Next step, is to work through some tutorials.

One thing to note is that the links to the slides are broken in the docs, but there are available in the source tree.

0

I think I am understanding it a bit better. I kind of like the Application and TaskCollection object models. Still not clear to me yet on how to structure workflows.

0

Ahh, Programmer Documentation is where I think the real insights will come from.

0

I don't think anything in GC3Apps part of gc3pie are applicable to work currently on our plate.

Perhaps, they might serve as examples as to how to write our own scripts.

0

For example, the computational chemistry program called GAMESS with the script.

0

So I am thinking the GC3Apps is a collection of scripts to run well known "pipelines", "algorithms", or other units of known calculations.

0

Having a bit of difficulty understanding the architecture, especially in the GC3Apps section.

I'm sure it will click eventually, but as of right now I am completely stumped and feel like a fool. :(

0

Been reading through the docs and so far it sounds like there are at least some promising ideas to use.

I like the notion of having auth and resources setup via a config file.

0

I am taking a look at gc3pie as something to help in building dynamic workflows of large job campaigns to be executed in grid environments such as SGE