Skip to content
This repository was archived by the owner on Apr 24, 2023. It is now read-only.

Commit

Permalink
create demand resources for fifo failures as well (#25)
Browse files Browse the repository at this point in the history
  • Loading branch information
onursatici authored Apr 11, 2019
1 parent 8233771 commit 2d9ca94
Showing 1 changed file with 7 additions and 4 deletions.
11 changes: 7 additions & 4 deletions internal/extender/resource.go
Original file line number Diff line number Diff line change
Expand Up @@ -210,20 +210,23 @@ func (s *SparkSchedulerExtender) selectDriverNode(ctx context.Context, driver *v
}
usages.Add(s.overheadComputer.GetOverhead(ctx, availableNodes))
availableResources := resources.AvailableForNodes(availableNodes, usages)
applicationResources, err := sparkResources(ctx, driver)
if err != nil {
return "", failureInternal, werror.Wrap(err, "failed to get spark resources")
}
if s.isFIFO {
queuedDrivers, err := s.podLister.ListEarlierDrivers(driver)
if err != nil {
return "", failureInternal, werror.Wrap(err, "failed to list earlier drivers")
}
ok := s.fitEarlierDrivers(ctx, queuedDrivers, driverNodeNames, executorNodeNames, availableResources)
if !ok {
if err := s.createDemandForApplication(ctx, driver, applicationResources); err != nil {
return "", failureInternal, werror.Wrap(err, "earlier drivers do not fit to the cluster, but failed to create demand resource")
}
return "", failureEarlierDriver, werror.Error("earlier drivers do not fit to the cluster")
}
}
applicationResources, err := sparkResources(ctx, driver)
if err != nil {
return "", failureInternal, werror.Wrap(err, "failed to get spark resources")
}
driverNode, executorNodes, hasCapacity := s.binpacker.BinpackFunc(
ctx,
applicationResources.driverResources,
Expand Down

0 comments on commit 2d9ca94

Please sign in to comment.