Thanks for attaching the problem!
It has roughly 30000 propositions and 439851 actions, which is large,
but not impossibly large. It may be the case that the translation step
is less efficient than it might otherwise be because of negative
preconditions, for which some parts of the translator can use
exponential transformations. So it might help to get rid of negative
preconditions in the precompilation step that you are using.
But looking at the output, I'm not sure this would help; I think the
main reason for the runtime is just that the problem is large. On my
machine I managed to solve it in roughly 13 minutes using this
command:
./fast-downward.py --alias lama-first \
comp_preprocessed-storage_13_dom.pddl \
comp_preprocessed-storage_13_prob.pddl
This only runs the first stage of LAMA ("lama-first"), which does not
care about solution quality, so the solution quality is probably poor:
the total cost is 2061, and there are lots of "forgo" actions, which I
assume is bad. But at least it does find a solution.
[In the meantime, I have also tried the full LAMA configuration, using
"seq-sat-lama-2011" in the command above instead of "lama-first".
Running the search component for 30 minutes total, solutions of the
following cost are found: 2061, 2060, 2052, 2036, 2028, 1884, 1744,
1650. Considering how much time is wasted finding poor-quality
solutions, this is probably an instance where the LAMA configuration's
Weighted-A*-style search is not the best fit.]
I have attached the planner log. If you didn't solve it, maybe you ran
out of main memory and the translator started swapping? On my machine,
the translation component used roughly 8.6 GB of peak memory.
Looking at the log, the translation component takes roughly 8 minutes,
but no individual step stands out. It just takes long because the
translator is implemented in Python, which is not very fast, and the
problem is large. I think if the same code were implemented in C++, we
might be able to do the same processing in around a minute, but
perhaps not much faster than that.
The search component took roughly 5 minutes, of which the largest part
(roughly 3:30 minutes) was generating the successor generator, the
next was generating the landmarks and landmark orders (roughly 1:20
minutes), search took 3.5 seconds, and the remaining seconds were
various things such as parsing.
I think with algorithms that are better adapted to this kind of input,
we might perform the successor generator computation much faster.
Without looking into things much more deeply, I'm not sure if we could
make the landmark generation much faster.
Things that could be done to reduce the runtime and memory usage:
1) Don't use the Python translator.
One alternative is that you could generate the output.sas file that
the translator generates directly. For a pregrounded input, this
would not actually be very challenging.
If you are interested, we also have an unsupported special-purpose
C++ variant of the translator that can only deal with pregrounded
problems and has many limitations, but might work for your input.
Let me know if you want to look into this.
A third alternative is to reimplement the complete translator in
C++. This is something we might end up doing as a long-term goal,
but even if we do, it will likely take several years to complete.
2) Try to improve the successor generator generation.
This is only worth doing once the translator is no longer the
bottleneck, but I'm hopeful that with some work, this could be
improved significantly. At least I don't see why it should take so
long with a problem of this size.
3) Try to improve the landmark graph generation.
Here I cannot really say how feasible this would be without
understanding the problem a lot better. But in any case, this would
only be worth doing after the previous two points are resolved.
For now, I will mark this as resolved because I think there is nothing
we can immediately do here, but if you want, of course please feel
free to continue the discussion.
In the medium term, I'd be interested in collecting information about
large problem instances where we might be able to improve performance
with a concerted effort. Improving the code in a meaningful way
requires many examples to make sure we're actually using good
algorithms rather than overfitting to specific problems, so making
progress on this will take some time. For this purpose, I'd like to
add the problem and some information about it to a public repository
of "challenge problems". Would this be OK for you?
Can we also include your contact data, so that whoever works on this
in the future knows how to get in touch with you if they come up with
good improvements or if they have questions about the problem?
|