mirror of
https://github.com/pytorch/pytorch.git
synced 2025-10-21 05:34:18 +08:00
Summary: better safe than sorry. will throw if memory overlap detected when using planned tensors and debug mode is enabled -- this will make our planning unit tests more robust. Test Plan: ci Rollback Plan: Differential Revision: D77327841 Pull Request resolved: https://github.com/pytorch/pytorch/pull/157290 Approved by: https://github.com/SherlockNoMad, https://github.com/zhxchen17
41 lines
1.4 KiB
C++
41 lines
1.4 KiB
C++
#include <torch/nativert/executor/ExecutionPlanner.h>
|
|
#include <torch/nativert/executor/ExecutorConfig.h>
|
|
#include <torch/nativert/executor/SerialGraphExecutor.h>
|
|
|
|
namespace torch::nativert {
|
|
|
|
std::vector<c10::IValue> SerialGraphExecutor::execute(
|
|
ExecutionFrame& executionFrame,
|
|
std::vector<c10::IValue> inputs) {
|
|
fillUserInputs(executionFrame, std::move(inputs));
|
|
|
|
return executeWithPrefilledFrame(executionFrame);
|
|
}
|
|
|
|
std::vector<c10::IValue> SerialGraphExecutor::executeWithPrefilledFrame(
|
|
ExecutionFrame& executionFrame) {
|
|
executionFrame.withManagedMemory([&](const LayoutManager* layout_manager) {
|
|
// Execute kernels for all nodes except prim.Input and prim.Output
|
|
for (NodeIndex nodeIdx = 1; nodeIdx < nodeKernels_.size() - 1; ++nodeIdx) {
|
|
nodeKernels_[nodeIdx]->compute(executionFrame);
|
|
|
|
#ifndef NDEBUG
|
|
if (layout_manager != nullptr) {
|
|
layout_manager->assert_no_overlapping_storages(nodeIdx);
|
|
}
|
|
#endif
|
|
|
|
// don't free intermediate values when static memory planning is enabled
|
|
if (executorConfig_.tryFreeUnmanagedValuesAfterUse) {
|
|
// Free the intermediate values that are no used anymore
|
|
for (const auto& valueKey : execPlan_->valuesToFree[nodeIdx]) {
|
|
executionFrame.releaseValueIfNeeded(valueKey);
|
|
}
|
|
}
|
|
}
|
|
});
|
|
return executionFrame.tryMoveUserOutputs();
|
|
}
|
|
|
|
} // namespace torch::nativert
|