Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[WIP] Distributed processing #60

Open
wants to merge 11 commits into
base: master
Choose a base branch
from
Open

[WIP] Distributed processing #60

wants to merge 11 commits into from

Conversation

thesamovar
Copy link
Contributor

No description provided.

@thesamovar
Copy link
Contributor Author

@nippoo @rossant I have now implemented distributed processing using IPython.parallel in this branch. It won't be highly efficient yet because I did some slightly lame things in order to get it working at all, which should be possible to make more efficient. Also, try_splits is not distributed yet, and so this isn't at all faster. However, it should hopefully work for normal iterations now, and I think even with the inefficiencies for a large data set it will be much faster when distributed across multiple machines. Could you try this out? The ipython notebook in dev can be used, you need to start up an appropriate engine before running it. It may also be the case that KK2 needs to be installed on each machine it runs on? Not sure how IPython handles that.

@rossant
Copy link
Member

rossant commented Jun 14, 2015

thanks @thesamovar we'll have a look! yeah KK2 probably needs to be installed on every machine first. I'm wondering if we can use docker to simplify deployement on many machines -- there should not be a significant performance hit, but that remains to be checked

@nippoo
Copy link
Contributor

nippoo commented Jun 16, 2015

@rossant there's no need to build images or deploy them: assuming the hardware is the same, just share a folder in the PATH over NFS or similar. This is exactly how I share KK1/SD2 over our servers. Legion and every computing cluster will run the same software on all nodes and share a directory over NFS. It's simple!

@rossant
Copy link
Member

rossant commented Jun 16, 2015

@nippoo neat!

@rossant
Copy link
Member

rossant commented Jun 16, 2015

@nippoo wait, this will only work for our code, not for conda dependencies..?

@nippoo
Copy link
Contributor

nippoo commented Jun 16, 2015

It works for everything - just make sure the shared miniconda installation / venv is in the PATH...

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants