data partition within group#1
Merged
asfgit merged 3 commits intoapache:masterfrom May 17, 2015
Merged
Conversation
TODO 1. update the performance collection by reporting performance to the stub. 2. let workers pass requests to the stub without copying data (passing addr or param id). messages to servers are then generated by the stub which can aggregate gradients of shared parameters from all workers and collect the updated parameters for them.
…implify the logics. now workers send simple messages to the stub thread which construct the real update/get/put requests. the stub thread also handles the responses from servers. E.g., the get/update response is handled by the stub now. the workers then wait until its param's version is updated in the collect function. avoid deadlocks for param_dealer_ and layer_dealer_ 2. tested data partition in single group in one procs. 3. generate a json file under workspace/visualization representing the neural net structure. users can create an image using the python script (scirpt/graph.py) reading the json file.
asfgit
pushed a commit
that referenced
this pull request
Aug 30, 2016
check build python package in mac
nudles
pushed a commit
that referenced
this pull request
Nov 27, 2019
changes EXPECT_EQ to EXPECT_NEAR
nudles
pushed a commit
that referenced
this pull request
Aug 12, 2020
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
This training scheme partitions one batch data into sub-batches where each worker in the group process one sub-batch. It is implemented by partitioning the layers (except the data and parser layers) of the original neural network into sub-layers where each sub-layer holds a sub-batch of the features. These sub-layers share the same set of parameter objects. There is one worker who is the owner for each parameter object. Each worker owns a partition of the neural network and computes the gradients of parameters over it.
The work flow is,
TODO,