File tree Expand file tree Collapse file tree 1 file changed +3
-3
lines changed
torchrec/inference/inference_legacy/src Expand file tree Collapse file tree 1 file changed +3
-3
lines changed Original file line number Diff line number Diff line change @@ -105,9 +105,9 @@ GPUExecutor::GPUExecutor(
105105 if (gcConfig_->optimizationEnabled ) {
106106 gcConfig_->threadIdToNumForwards [threadId] = 0 ;
107107 // Freeze all python objects in each interpreter
108- auto model =
108+ auto session =
109109 model_.acquireSession (&manager_->allInstances ().at (threadId));
110- model .global (" gc" , " freeze" )(at::ArrayRef<torch::deploy::Obj>());
110+ session .global (" gc" , " freeze" )(at::ArrayRef<torch::deploy::Obj>());
111111 }
112112
113113 processThreads_.emplace_back ([this , threadId] {
@@ -189,7 +189,7 @@ void GPUExecutor::process(int idx) {
189189 auto start = std::chrono::steady_clock::now ();
190190 model_.acquireSession (&manager_->allInstances ().at (idx));
191191 {
192- std::lock_guard<std::mutex> lock (warmUpMutex_);
192+ std::lock_guard<std::mutex> warmUpLock (warmUpMutex_);
193193 warmUpCounter_++;
194194 warmUpCV_.notify_one ();
195195 }
You can’t perform that action at this time.
0 commit comments