Files
pytorch/torch/csrc/distributed/rpc/request_callback.cpp

22 lines
832 B
C++

#include <torch/csrc/distributed/rpc/request_callback.h>
#include <torch/csrc/distributed/autograd/context/container.h>
#include <torch/csrc/distributed/autograd/utils.h>
namespace torch::distributed::rpc {
using namespace torch::distributed::autograd;
c10::intrusive_ptr<JitFuture> RequestCallback::operator()(
Message& request,
std::vector<c10::Stream> streams) const {
// NB: cannot clear autograd context id here because the processMessage method
// might pause waiting for all RRefs in the arguments to be confirmed by their
// owners and resume processing in a different thread. Hence, the
// thread_local context id needs to be set and cleared in the thread that
// indeed carries out the processing logic.
return processMessage(request, std::move(streams));
}
} // namespace torch::distributed::rpc