83 Matching Annotations
  1. Last 7 days
  2. Aug 2025
  3. May 2025
  4. Apr 2025
    1. zkt + T k(zt, at), zkt+1

      Won't this always be 0 since the first and second term in the distance function is calculated in the same way? The definition of z_t+1 (as defined earlier) is "z_t+1 = z_t + T(z_t,a_t)"

    2. Each feature map mkt = [Eext(st)]k can be interpreted as an object mask correspondingto one particular object slot, where [. . .]k denotes selection of the k-th feature map

      How do they make sure that each feature map actually represents an object, and not something else from the image (e.g. just a fixed section of the image)?

  5. Mar 2025
  6. Feb 2025
    1. On each iteration, the single predicate that improves Jsurr the most is added to the se

      What if none of the invented predicates makes the problem solvable? Then every predicate should have the same evaluation?

    1. Forthe Hanoi domain, L is identical through all problems.

      This NL instruction is actually interesting, and perhaps quite complex for the model to handle. But if the instruction is the same across all tasks, it can essentially copy the prompt examples and use that PDDL.

    2. The three pegs arenamed by the number from left to right (e.g., peg1, peg2,and peg3)

      What if the pegs would not be named in order? Would that reduce performance? Now the model can rely on the object name index to determine the order.

  7. Jan 2025
    1. All four PDDL problem descriptions represent the planning problem of stacking one blockonto another

      However, all goal does not reflect the goal state shown in the image, only the first do

  8. Dec 2024
    1. Somewhat surprisingly, the lowest scores in Blocksworldare associated with BlockAmbiguity and KStacksColor; thesetwo problems require the LLM to associate objects basedon their color and we had apriori expected the LLM to becapable of such associations and perform well on this task.

      This kind of makes sense, because the colors are not explicitly modelled but they are part of the make of the object e.g. "red_block_1" rather than "red(block_1)". The latter would be a more natural way to express colors, as colors is a property of an object

  9. Nov 2024
  10. Jul 2024
  11. Jun 2024
  12. May 2024
  13. Feb 2024
    1. One can view the noise vector z in such a GAN as a featurevector, containing some representation of the transition to o′ from o.

      How can it contain a representation of the transition if it is just noise?

  14. Jan 2024
    1. showing good variability over the irrelevant factors

      Not really. For the "white suitcase" scene it only differs in wall colors and floor colors, but the "black and white" representation of the scene is the same. Essentially there could be a way larger range of scenes where a white suitcase appears.

    1. In one case, the input data corresponds to one or morestate graphs Gi assumed to originate from hidden planninginstances Pi = 〈D, Ii〉 that need to be uncovered

      Isn't the domain needed in order to generate the state graph?

    1. prior distribution over programs likely to solve tasks inthe domain

      What does this prior distribution mean? The probability of the program to solve any task in the domain? Is there even any programs that would solve multiple tasks?

  15. Dec 2023
    1. Task descriptions are constructed using PDDL and symbolicplans are generated using the FAST-DOWNWARD planner

      To generate a symbolic plan, an initial state (problem file) needs to be given. How does this looks like? Is there only three problem files (one for each problem) representing some "general" state? Shouldn't the initial plan depend on the initial state?

    1. Ourset-up automatically parses LLM-generated language intoa program using our synthetic grammar

      How?

      Also, how do they handle cases where the parser generates incorrect PDDL? Wouldn't that give the LLM-as-planner a worse score that it actually should have?

    1. Even with high Exec, some task GCR are low, becausesome tasks have multiple appropriate goal states, but weonly evaluate against a single “true” goal

      This seems like an unfair way to evaluate the model

  16. Nov 2023
    1. Generated strings from all models arepost-processed for common errors in sequence-to-sequence models, including token doubling,completing missing bigrams (e.g. “pick <arg1>”→ “pick up <arg1>”), and heuristics for addingmissing argument tags

      Probably won't generalize well to new domains

    1. hus, we have created a PDDL domain file usingour knowledge of the objects and actions in the ALFRED world and a PDDL problem file for eachsample

      I assume that the domain file is created manually, but are the problem files also created by hand? If so that seems like a lot of work, since the dataset has 8,055 visual samples, the same amount would be needed to be handcoded.

    2. Since in our task we ignorethe vision part of the data, we might encounter some duplicates between our datasets

      How do they get the scene representation from the visual data? Is this included in the ALFRED dataset?

  17. Oct 2023
    1. ALFWorld uses PDDL - Planning DomainDefinition Language (McDermott et al., 1998) to describe each scene from ALFRED and to constructan equivalent text game using the TextWorld engine.

      How is the PDDL created?