This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

FoREST: Frame of Reference Evaluation in Spatial Reasoning Tasks

Tanawan Premsri
Department of Computer Science
Michigan State University
premsrit@msu.edu
\AndParisa Kordjamshidi
Department of Computer Science
Michigan State University
kordjams@msu.edu
Abstract

Spatial reasoning is a fundamental aspect of human intelligence. One key concept in spatial cognition is the Frame of Reference (FoR), which identifies the perspective of spatial expressions. Despite its significance, FoR has received limited attention in AI models that need spatial intelligence. There is a lack of dedicated benchmarks and in-depth evaluation of large language models (LLMs) in this area. To address this issue, we introduce the Frame of Reference Evaluation in Spatial Reasoning Tasks (FoREST) benchmark, designed to assess FoR comprehension in LLMs. We evaluate LLMs on answering questions that require FoR comprehension and layout generation in text-to-image models using FoREST. Our results reveal a notable performance gap across different FoR classes in various LLMs, affecting their ability to generate accurate layouts for text-to-image generation. This highlights critical shortcomings in FoR comprehension. To improve FoR understanding, we propose Spatial-Guided prompting, which improves LLMs’ ability to extract essential spatial concepts. Our proposed method improves overall performance across spatial reasoning tasks.

FoREST: Frame of Reference Evaluation in Spatial Reasoning Tasks


Tanawan Premsri Department of Computer Science Michigan State University premsrit@msu.edu                        Parisa Kordjamshidi Department of Computer Science Michigan State University kordjams@msu.edu


1 Introduction

Spatial reasoning plays a significant role in human cognition and daily activities. It is also a crucial aspect in many AI problems, including language grounding (Zhang and Kordjamshidi, 2022; Yang et al., 2024), navigation (Yamada et al., 2024), computer vision (Liu et al., 2023; Chen et al., 2024), medical domain (Gong et al., 2024), and image generation (Gokhale et al., 2023). One key concept in spatial reasoning is the Frame of Reference (FoR), which identifies the perspective of spatial expressions. FoR has been studied extensively in cognitive linguistics (Edmonds-Wathen, 2012; Vukovic and Williams, 2015). Levinson (2003) initially defines three FoR classes: relative, based on the observer’s perspective; intrinsic, based on an inherent feature of the reference object; and absolute, using environmental cues like cardinal directions -See Figure 1. This framework was expanded by  Tenbrink (2011) to create a more comprehensive framework, serving as the basis of our work. Understanding FoR is important for many applications, especially in embodied AI. In such applications, an agent must simultaneously comprehend multiple perspectives, including the one from the instruction giver and from the instruction follower, to communicate and perform tasks effectively. However, recent spatial evaluation benchmarks have largely overlooked FoR. For example, the text-based benchmarks Shi et al. (2022); Mirzaee and Kordjamshidi (2022); Rizvi et al. (2024) and text-to-images benchmarks (Gokhale et al., 2023; Huang et al., 2023; Cho et al., 2023a, b) assume a fixed perspective for all spatial expressions. This inherent bias limits situated spatial reasoning, restricting adaptability in interactive environments where perspectives can change.

Refer to caption
Figure 1: Illustration of FoR classes. The cat is the locatum, the car is the relatum, and the arrow indicates the perspective.

To systematically investigate the role of FoR in spatial understanding and create a new resource, that is, Frame of Reference Evaluation in Spatial Reasoning Tasks (FoREST), FoREST is designed to evaluate models’ ability to comprehend FoR from textual descriptions and extend this evaluation to grounding and visualization. Our benchmark includes spatial expressions with FoR ambiguity—where multiple FoRs may apply to the described situation—and spatial expressions with only a single valid FoR. This design allows evaluation of the models’ understanding of FoR in both scenarios. We evaluate several LLMs in a QA setting that require FoR understanding and apply the FoR concept in text-to-image models. Our findings reveal performance differences across FoR classes and show that LLMs exhibit bias toward specific FoRs when handling ambiguous cases. This bias extends to layout-diffusion models, which rely on LLM-generated layouts in the image generation pipeline. To enhance FoR comprehension in LLMs, we propose Spatial-Guided prompting, which enables models to analyze and extract additional spatial information, including directional, topological, and distance relations. We demonstrate that incorporating spatial information improves question-answering and layout generation, ultimately enhancing text-to-image generation performance.

Our contribution111code and dataset available at GitHub repository. are summarized as follows, 1. We introduce the FoREST benchmark to systematically evaluate LLMs’ FoR comprehension in a QA setting. 2. We analyze the impact of FoR information on text-to-image generation using multiple diffusion models. 3. We propose a prompting approach that generates spatial information, which can be incorporated into QA and layout diffusion to enhance performance.

2 Spatial Primitives

We review three semantic aspects of spatial information expressed in language: Spatial Roles, Spatial Relations, and Frame of Reference.

Spatial Roles. We focus on two main spatial roles (Kordjamshidi et al., 2010) of Locatum, and Relatum. The locatum is the object described in the spatial expression, while the relatum is the other object used to describe the position of the locatum. An example is a cat is to the left of a dog, where the cat is the locatum, and the dog is the relatum.

Spatial Relations. When dealing with spatial knowledge representation and reasoning, three main relations categories are often considered, that is, directional, topological, and distance (Hernández, 1994; Cohn and Renz, 2008; Kordjamshidi et al., 2011). Directional describes an object’s direction based on specific coordinates. Examples of relations include left and right. Topological describes the containment between two objects, such as inside. Distance describes qualitative and quantitative relations between entities. Examples of qualitative are far, and quantitative are 3km.

Spatial Frame of Reference. We use four frames of references investigated in the cognitive linguistic studies Tenbrink (2011). These are defined based on the concept of Perspective, which is the origin of a coordinate system to determine the direction. The four frames of reference are defined as follows.

1. External Intrinsic describes a spatial relation from the relatum’s perspective, where the relatum does not contain the locatum. The top-right image in Figure 1 illustrates this with the sentence, A cat is to the right of the car from the car’s perspective.

2. External Relative describes a spatial relation from the observer’s perspective. The top-left image in Figure 1 shows an example with the sentence, A cat is to the left of a car from my perspective.

3. Internal Intrinsic describes a spatial relation from the relatum’s perspective, where the relatum contains the locatum. The bottom-right image in Figure 1 show this with the sentence, A cat is inside and back of the car from the car’s perspective.

4. Internal Relative describes a spatial relation from the observer’s perspective where the locatum is inside the relatum. The bottom-left image in Figure 1 show this FoR with the sentence, A cat is inside and to the left of the car from my perspective.

Refer to caption
Figure 2: Pipeline for dataset creation, starting from selecting a locatum and relatum from available objects and then applying a spatial template to generate the spatial expression (TT). FoRs are assigned based on the relatum’s properties. TT is then categorized based on the number of FoRs. For example, A cat is to the right of a dog (with two possible FoRs: external intrinsic and external relative) belongs to the A-split. Then, its disambiguated version (A cat is to the right of a dog from the dog’s perspective) is added to the C-split. Next, if applicable, a relatum’s orientation is included for visualization and question generation. Finally, Unity3D generates scene configurations, and question-answer pairs are created from TT.

3 FoREST Dataset Construction

To systematically evaluate LLM on the frame of reference (FoR) recognition, we introduce the Frame of Reference Evaluation in Spatial Reasoning Tasks (FoREST) benchmark. Each instance in FoREST consists of a spatial context (TT), a set of corresponding FoR (FoRFoR) which is a subset of {external relative, external intrinsic, internal intrinsic, internal relative}, a set of questions and answers ({Q,A}\{Q,A\}), and a set of visualizations ({I}\{I\}). An example of TT is A cat is to the right of a dog. A dog is facing toward the camera. The FoR of this expression is {external intrinsic, external relative}. A possible question-answer is QQ = Based on the camera’s perspective, where is the cat from the dog’s position?, AA = {left, right}. There is an ambiguity in the FoR for this expression. Thus, the answer will be left if the model assumes the external relative. Conversely, it will be right if the model assumes the external intrinsic. The visualization of this example is in Figure 2.

3.1 Context Generation

We select two distinct objects—a relatum (RR) and a locatum (LL)—from a set of 20 objects and apply them to a Spatial Relation template, <LL> <spatial relation> <RR> to generate the context TT. FoRs for TT are determined based on the properties of the selected objects. Depending on the number of possible FoRs, TT is categorized as ambiguous (A-split), where multiple FoRs apply, or clear (C-split), where only one FoR is valid. We further augment the C-split with disambiguated spatial expressions derived from the A-split, as shown in Figure 2.

3.2 Categories based on Relatum Properties

Using the FoR classes in Section 2, we identified two key relatum properties contributing to FoR ambiguity. The first property is the relatum’s intrinsic direction. It creates ambiguity between intrinsic and relative FoR since spatial relations can originate from the relatum’s and observer’s perspectives. The second is the relatum’s affordance as a container. It introduces the ambiguity between internal and external FoR, as spatial relations can refer to the inside and outside of the relatum. Based on these properties, we define four distinct cases: Cow Case, Box Case, Car Case, and Pen Case.

Case 1: Cow Case. In this case, the selected relatum has intrinsic directions but does not have the affordance as the container for the locatum. An obvious example is a cow, which should not be a container but has a front and back. In such cases, the relatum potentially provides a perspective for spatial relations. The applicable FoR classes are FoRFoR = {external intrinsic, external relative}. We augment the C-split with expressions of this case but include the perspective to resolve their ambiguity. To specify the perspective, we use predefined templates for augmenting clauses, such as from {relatum}’s perspective for external intrinsic or from the camera’s perspective for external relative. For example, if the context is A cat is to the right of the cow, in the A-split. The counterparts included in the C-split are A cat is to the right of the cow from cow’s perspective. for external intrinsic and A cat is to the right of the cow from my perspective. for external intrinsic.

Case 2: Box Case. The relatum in this category has the property of being a container but lacks intrinsic directions, making the internal FoR applicable. An example is a box. The applicable FoR classes are FoRFoR = {external relative, internal relative}. To include their unambiguous counterparts in the C-split, we specify the topological relation to the expressions, TT, by adding inside for internal relative and outside for external relative cases. For example, for the sentence A cat is to the right of the box., the unambiguous TT with internal relative FoR is A cat is inside and to the right of the box. The counterpart for external relative is A cat is outside and to the right of the box.

Case 3: Car Case. A relatum with an intrinsic direction and container affordance falls into this case, allowing all FoR classes. An obvious example is a car that can be a container with intrinsic directions. The applicable FoR classes are FoRFoR = { external relative, external intrinsic, internal intrinsic, internal relative}. To augment C-split with this case’s disambiguated counterparts, we add perspective and topology information similar to the Cow and Box cases. An example expression for this case is A person is in front of the car. The four disambiguated counterparts to include in the C-split are A person is outside and in front of the car from the car itself. for external intrinsic, A person is outside and in front of the car from the observer. for external relative, A person is inside and in front of the car from the car itself. for internal intrinsic, and A person is inside and in front of the car from the observer. for internal relative.

Case 4: Pen Case. In this case, the relatum lacks both the intrinsic direction and the affordance as a container. An obvious example is a pen with neither left/right nor the ability to be a container. Lacking these two properties, the created context has only one applicable FoR, FoRFoR = { external relative}. Therefore, we can categorize this case into both splits without any modification. An example of such a context is The book is to the left of a pen.

3.3 Context Visualization

In our visualization, complexity arises when the relatum has an intrinsic direction within the intrinsic FoR, as its orientation can complicate the spatial representation. For example, for visualizing A cat is to the right of a dog from the dog’s perspective., the cat can be placed in different coordinates based on the dog’s orientation. To address this issue, we add a template sentence for each direction, such as <relatum> is facing toward the camera, to specify the relatum’s orientation of all applicable TT for visualization and QA. For instance, A cat is to the left of a dog. becomes A cat is to the left of a dog. The dog is facing toward the camera. To avoid occlusion issues, we generate visualizations only for external FoRs, as one object may become invisible in internal FoR classes. We use only expressions in C-split since those have a unique FoR interpretation for visualization. We then create a scene configuration by applying a predefined template, as illustrated in Figure 2. Images are generated using the Unity 3D simulator Juliani et al. (2020), producing four variations per expression TT with different backgrounds and object positions. Further details on the simulation process are in Appendix B.

3.4 Question-Answering Generation

We generate questions for all generated spatial expressions (TT). Note that we include the relatum orientation for cases where the relatum has an intrinsic direction, as mentioned in the visualization. Our benchmark includes two types of questions. The first type asks for the spatial relation between two given objects from the camera’s perspective, following predefined templates such as, Based on the camera’s perspective, where is the locatum relative to the relatum’s position? Template variations are made based on GPT4o. The second type of question queries the spatial relation from the relatum’s perspective. This question type follows the same templates but replaces the camera with the relatum. The first type of question is generated for all TT, while the second type is only generated for TT where the relatum has intrinsic direction and a perspective can be defined accordingly. Question templates are provided in Appendix B.3. Answers are determined based on the corresponding FoRs, the spatial relation in TT, and the relatum’s orientation when applicable.

4 Models and Tasks

The FoREST benchmark supports multiple tasks, including FoR identification, Question Answering (QA) that requires FoR comprehension, and Text-to-Image (T2I). This paper focuses on QA and T2I for a deeper evaluation of spatial reasoning. FoR identification experiments are provided in Appendix E.

4.1 Question-Answering (QA)

Task. This QA task evaluates LLMs’ ability to adapt contextual perspectives across different FoRs. Both A and C splits are used in this task. The input is the context, consisting of a spatial expression TT and relatum orientation, if available, and a question QQ that queries the spatial relation from either an observer or the relatum’s perspective. The output is a spatial relation SS, restricted to {left, right, front, back}.

Zero-shot baseline. We call the LLM with instructions, a spatial context, TT, and a question, QQ, expecting a spatial relation as the response. The prompt instructs the model to answer the question with one of the candidate spatial relations without any explanations.

Few-shot baseline. We create four spatial expressions, each assigned to a single FoR class to prevent bias. Following the steps in Section 3.4, we generate a corresponding question and answer for each. These serve as examples in our few-shot prompting. The input to the model is instruction, example, spatial context, and the question.

Chain-of-Thought baseline (Wei et al., 2023). To create Chain-of-Thought (CoT) examples, we modify the prompt to require reasoning before answering. We manually crafted reasoning explanations with the necessary information for each example we used in the few-shot setting. The input to the model is instruction, CoT example, spatial context, and the question.

Camera perspective Relatum perspective
Model Cow Car Box Pen Avg. Cow Car Avg.
R% I% Acc. R% I% Acc. Acc. Acc. Acc. R% I% Acc. R% I% Acc. Acc.
Llama3-70B (1) 48.148.1 51.5\mathbf{51.5} 62.562.5 58.0\mathbf{58.0} 41.641.6 65.565.5 73.373.3 72.572.5 64.364.3 61.0\mathbf{61.0} 38.738.7 62.162.1 51.8\mathbf{51.8} 47.947.9 61.861.8 62.162.1
Llama3-70B (2) 49.149.1 50.5\mathbf{50.5} 62.262.2 52.2\mathbf{52.2} 47.447.4 64.764.7 85.885.8 85.585.5 65.865.8 59.6\mathbf{59.6} 40.140.1 57.157.1 55.5\mathbf{55.5} 44.244.2 61.861.8 57.757.7
Llama3-70B (3) 49.449.4 50.3\mathbf{50.3} 80.780.7 49.449.4 50.3\mathbf{50.3} 79.679.6 95.895.8 94.994.9 82.682.6 60.8\mathbf{60.8} 39.039.0 77.277.2 55.1\mathbf{55.1} 44.644.6 80.980.9 77.777.7
Llama3-70B (4) 59.4\mathbf{59.4} 40.240.2 73.673.6 57.9\mathbf{57.9} 41.741.7 74.874.8 100.0100.0 100.0100.0 77.577.5 60.6\mathbf{60.6} 39.139.1 65.765.7 56.0\mathbf{56.0} 43.743.7 67.767.7 66.066.0
Qwen2-72B (1) 96.6\mathbf{96.6} 2.92.9 95.695.6 95.9\mathbf{95.9} 3.63.6 95.095.0 100.0100.0 100.0100.0 96.196.1 8.88.8 90.6\mathbf{90.6} 79.379.3 7.87.8 91.7\mathbf{91.7} 83.683.6 79.979.9
Qwen2-72B (2) 89.0\mathbf{89.0} 10.510.5 84.484.4 85.6\mathbf{85.6} 13.913.9 85.585.5 100.0100.0 100.0100.0 86.886.8 17.717.7 81.8\mathbf{81.8} 78.378.3 10.410.4 89.1\mathbf{89.1} 86.386.3 79.479.4
Qwen2-72B (3) 67.2\mathbf{67.2} 32.432.4 88.688.6 62.0\mathbf{62.0} 37.637.6 83.483.4 100.0100.0 100.0100.0 89.689.6 21.321.3 78.3\mathbf{78.3} 85.585.5 22.722.7 76.9\mathbf{76.9} 83.683.6 85.285.2
Qwen2-72B (4) 93.0\mathbf{93.0} 6.56.5 90.190.1 94.6\mathbf{94.6} 4.94.9 93.393.3 100.0100.0 98.698.6 91.791.7 8.28.2 91.2\mathbf{91.2} 86.086.0 10.510.5 89.0\mathbf{89.0} 87.487.4 86.286.2
GPT-4o (1) 84.3\mathbf{84.3} 15.315.3 94.594.5 88.5\mathbf{88.5} 11.011.0 97.397.3 99.299.2 99.899.8 95.695.6 21.621.6 78.0\mathbf{78.0} 91.691.6 16.116.1 83.5\mathbf{83.5} 90.590.5 91.491.4
GPT-4o (2) 69.0\mathbf{69.0} 30.630.6 76.676.6 80.3\mathbf{80.3} 19.219.2 89.589.5 100.0100.0 100.0100.0 81.581.5 29.029.0 70.5\mathbf{70.5} 74.774.7 30.930.9 68.7\mathbf{68.7} 77.577.5 75.175.1
GPT-4o (3) 41.541.5 58.3\mathbf{58.3} 92.392.3 38.238.2 61.6\mathbf{61.6} 91.091.0 100.0100.0 99.899.8 93.293.2 33.933.9 65.8\mathbf{65.8} 93.993.9 32.032.0 67.6\mathbf{67.6} 93.993.9 93.993.9
GPT-4o(4) 26.026.0 73.9\mathbf{73.9} 79.279.2 27.727.7 72.1\mathbf{72.1} 79.479.4 96.796.7 94.394.3 81.481.4 16.216.2 83.4\mathbf{83.4} 95.595.5 19.219.2 80.4\mathbf{80.4} 94.894.8 95.495.4
Table 1: QA accuracy in the A-Split across various LLMs. R% and I% represent the percentage the model assumes relative or intrinsic FoR for ambiguous expression explained in Section 5.1. Acc is the accuracy, and Avg is the micro-average of accuracy. (1): 0-shot, (2): 4-shot, (3): CoT, and (4): SG + CoT.

4.2 Text-To-Image (T2I)

Task. This task aims to determine the diffusion models’ ability to consider FoR by evaluating their generated images. The input is a spatial expression, TT, and the output is a generated image (II). We use the context from both C and A splits with external FoRs for this task.

Stable Diffusion Models. We use the stable diffusion models as the baseline for the T2I task. This model only needs the scene description as input.

Layout Diffusion Models. We evaluate the Layout Diffusion model, a more advanced T2I model operating in two phases: text-to-layout and layout-to-image. Given that LLMs can generate the bounding box layout (Cho et al., 2023b), we provide them with instructions and TT to create the layout. The layout consists of bounding box coordinates for each object in the format of {object: [x,y,w,h][x,y,w,h]}, where xx and yy denote the starting point and hh and ww denote the height and width. The bounding box coordinates and TT are then passed to the layout-to-image model to produce the final image, II.

4.3 Spatial-Guide Prompting

We hypothesize that the spatial relation types and FoR classes explained in Section 2 can improve question-answering and layout generation. For instance, the external intrinsic FoR emphasizes that spatial relations originate from the relatum’s perspective. To leverage this, we propose Spatial-Guided (SG) prompting, an additional step applied before QA or layout generation steps. This step extracts spatial information, including direction, topology, distance as well as the FoR from spatial expression TT. The extracted information will serve as supplementary for guiding LLMs in QA and layout generation. We manually craft four examples covering these aspects. First, we specify the perspective for directional relations, e.g., left relative to the observer, to distinguish intrinsic from relative FoR. Next, we indicate whether the locatum is inside or outside the relatum for topological relations to differentiate internal from external FoR. Lastly, we provide an estimated quantitative distance to support topological and directional relation identification, such as far. These examples are then provided as a few-shot example for the model to extracted information automatically.

5 Experimental Results

Camera perspective Relatum perspective
Model ER (CP) EI (RP) II (RP) IR (CP) Avg. ER (CP) EI (RP) II (RP) IR (CP) Avg.
Llama3-70B (0-shot) 44.844.8 38.438.4 39.739.7 54.454.4 42.642.6 42.242.2 47.147.1 62.562.5 34.434.4 45.145.1
Llama3-70B (4-shot) 43.043.0 40.040.0 39.139.1 47.347.3 41.941.9 41.841.8 60.960.9 77.777.7 35.235.2 52.052.0
Llama3-70B (CoT) 57.857.8 46.146.1 44.744.7 46.046.0 51.551.5 55.5\mathbf{55.5} 56.856.8 71.571.5 49.049.0 56.656.6
Llama3-70B (SG + CoT) 47.647.6 42.942.9 50.050.0 35.635.6 45.045.0 55.455.4 64.564.5 75.075.0 47.147.1 60.160.1
Qwen2-72B (0-shot) 94.594.5 35.235.2 31.831.8 93.293.2 66.966.9 28.728.7 89.389.3 93.693.6 23.823.8 59.059.0
Qwen2-72B (4-shot) 90.290.2 39.539.5 39.139.1 68.568.5 65.365.3 33.533.5 92.192.1 94.094.0 29.529.5 62.762.7
Qwen2-72B (CoT) 81.481.4 57.457.4 58.658.6 62.562.5 69.169.1 39.539.5 83.783.7 85.285.2 37.737.7 61.661.6
Qwen2-72B (SG + CoT) 97.697.6 42.542.5 31.331.3 93.893.8 71.471.4 42.842.8 86.686.6 92.092.0 34.034.0 64.564.5
GPT-4o (0-shot) 79.779.7 45.145.1 39.539.5 90.290.2 64.264.2 46.946.9 88.588.5 98.298.2 34.834.8 67.567.5
GPT-4o (4-shot) 68.068.0 52.652.6 60.760.7 74.174.1 61.861.8 44.944.9 98.2\mathbf{98.2} 100.0\mathbf{100.0} 37.537.5 71.271.2
GPT-4o (CoT) 81.781.7 76.1\mathbf{76.1} 82.4\mathbf{82.4} 71.571.5 78.878.8 53.053.0 91.191.1 90.690.6 50.8\mathbf{50.8} 71.971.9
GPT-4o (SG + CoT) 97.9\mathbf{97.9} 72.272.2 72.772.7 93.4\mathbf{93.4} 85.8\mathbf{85.8} 48.948.9 96.396.3 95.995.9 36.136.1 71.8\mathbf{71.8}
Table 2: QA accuracy in the C-Split across various LLMs. ER, EI, II, and IR denote external relative, external intrinsic, internal intrinsic, and internal relative FoRs. Avg represents the micro-average accuracy. CP refers to context with camera perspective, while RP denotes context with relatum perspective.

5.1 Evaluation Metrics

QA. We report an accuracy measure defined as follows. Since the questions can have multiple correct answers, specifically in A-split, as explained in Section 3, the prediction is correct if it matches any valid answer. Additionally, we report the model’s bias distribution when FoR ambiguity exists. II% is the percentage of correct answers when assuming an intrinsic FoR, while RR% is this percentage with a relative FoR assumption. Note that cases where both FoR assumptions lead to the same answer are excluded from these calculations.

T2I. We adopt spatialEval (Cho et al., 2023b) approach for evaluating T2I spatial ability. However, we modify it to account for FoR. We convert all relations to a camera perspective before passing them to spatialEval, which assumes this viewpoint. Accuracy is determined by comparing the bounding box and depth map of the relatum and locatum. For FoR ambiguity, a generated image is correct if it aligns with at least one valid FoR interpretation. We report results using VISORcond and VISORuncond (Gokhale et al., 2023), metrics for assessing T2I spatial understanding. VISORcond evaluates spatial relations only when both objects appear correctly, aligning with our focus on spatial reasoning rather than object creation. In contrast, VISORuncond evaluates the overall performance, including object creation errors.

5.2 Experimental Setting

QA. We use Llama3-70B (Llama, 2024), Qwen2-72B (Qwen Team, 2024), and GPT-4o (gpt-4o-2024-11-20(OpenAI, 2024) as the backbones for prompt engineering. To ensure reproducibility, we set the temperature of all models to 0. For all models, we apply zero-shot, few-shot, CoT, and our proposed prompting with CoT (SG+CoT).

T2I. We select Stable Diffusion SD-1.5 and SD-2.1 (Rombach et al., 2021) as our stable diffusion models and GLIGEN(Li et al., 2023) as the layout-to-image backbone. For translating spatial descriptions into textual bounding box information, we use Llama3-8B and Llama3-70B, as detailed in Section 4.2. The same LLMs are used to generate spatial information for SG prompting. We generate four images to compute the VISOR score following Gokhale et al. (2023) Inference steps for all T2I models are set to 50. For the evaluation modules, we select grounding DINO (Liu et al., 2025) for object detection and DPT (Ranftl et al., 2021) for depth mapping, following VPEval Cho et al. (2023b). The experiments were conducted on an A6000A6000 GPU, totaling approximately 300300 GPU hours.

5.3 Results

RQ1. What is the bias of the LLMs for the ambiguous FoR? Table 1 presents the QA results for the A-split. Ideally, a model that correctly extracts the spatial relation without considering perspective should achieve 100% accuracy, as the context lacks a fixed perspective. However, this ideal model is not the focus of our work. We aim to assess model bias by measuring how often LLMs adopt a specific perspective when answering. In the Cow and Pen case, relatum properties do not introduce FoR ambiguity in directional relations, making the task pure extraction rather than reasoning. Thus, we focus on the II% and RR% of the Cow and Car cases, which best reflect LLMs’ bias. Qwen2 achieves around 80% accuracy across all experiments by selecting spatial expressions directly from the context, suggesting it may disregard the question’s perspective. GPT-4o shows similar bias in 0-shot and 4-shot settings but shifts toward intrinsic interpretation with CoT. This bias reduces accuracy in camera-perspective questions from 93.2% to 81.4%, where FoR adaptation is more challenging than relation extraction. Llama3-70B lacks a strong preference, balancing assumptions but slightly favoring relative FoR. This uncertainty lowers performance, requiring more reasoning to reach the correct answer. In summary, Qwen2 achieves higher accuracy by focusing on relation extraction without considering FoR reasoning, while other models attempt reasoning but struggle to reach correct conclusions, leading to lower performance.

VISOR(%)
A-Split C-Split
Model cond (I) cond (R) cond (avg) cond (I) cond (R) cond (avg)
EI FoR ER FoR all
SD-1.5 51.1151.11 21.6121.61 72.7272.72 48.9548.95 68.7268.72 53.9253.92 53.7753.77 53.8353.83
SD-2.1 57.9757.97 21.4921.49 79.4679.46 54.1054.10 75.3975.39 60.06\mathbf{60.06} 59.6459.64 59.8359.83
Llama3-8B + GLIGEN 53.6753.67 25.7825.78 79.4579.45 66.0866.08 77.3877.38 57.5157.51 65.9865.98 62.1262.12
Llama3-70B + GLIGEN 54.4954.49 29.4529.45 83.9483.94 68.6868.68 81.4381.43 56.4756.47 69.5369.53 63.4963.49
Llama3-8B + SG + GLIGEN (Our) 57.4657.46 27.9627.96 85.4285.42 71.14\mathbf{71.14} 83.1783.17 58.8458.84 70.36\mathbf{70.36} 65.15\mathbf{65.15}
Llama3-70B + SG + GLIGEN (Our) 56.5456.54 30.5930.59 87.13\mathbf{87.13} 66.5666.56 83.75\mathbf{83.75} 56.7756.77 70.04{70.04} 64.06{64.06}
Table 3: VISORcond score on the A and C splits where II refer to the Cow case and Car case where relatum has intrinsic directions, and RR refer to the Box case and Pen case where relatum lacks intrinsic directions, avgavg is mirco-average of II and RR. cond are explained in Section 5.1. EI and ER FoR represent the generated image considered corrected by EI or ER FoR
Refer to caption
Figure 3: Confusion matrices of spatial relation answers when Llama3 and GPT-4o are required to adapt FoR in the 0-shot and (SG+CoT) settings.

RQ2. Can the model adapt FoR when answering the questions? To address this research question, we analyze QA that required FoR comprehension results in C-Split from Table 2. Note that the context and question in these tasks explicitly indicate a perspective. The results indicate that LLMs struggle with FoR conversion, particularly when the question has relatum and the context has camera perspectives, achieving only up to 55.5% accuracy. We further demonstrate how Llama3 and GPT-4o adapt FoR using the confusion matrix in Figure 3. Our findings reveal that pure-text LLM (Llama3) has confusion between left and right. Humans typically reverse front and back while preserving left and right when describing the spatial relation from perspective. However, Llama3 incorrectly reverses left and right, leading to poor adaptation to the camera perspective. In contrast, very large multimodal-language models like GPT-4o follow the expected pattern, as observed by Zhang et al. 2025. While our GPT-4o results suggest some ability to convert the relatum’s perspective into the camera’s with in-context learning (72% accuracy), the reverse transformation in the textual domain remains challenging (53% accuracy). This difficulty persists when converting spatial relations from the camera perspective from images to the relatum’s perspective as observed in Zhang et al. 2025.

RQ3. How can an explicit FoR identification help spatial reasoning in QA? We compare CoT and CoT+SG results to evaluate how explicit FoR identification affects LLMs’ spatial reasoning in QA. Based on C-Split results (Table 3), incorporating SG encourages the model to identify the perspective from input expression ranging from 2.9% to 30% in cases where the context and question share the same perspective. These cases are easier as the models do not need FoR adaptation. With one exception, as can be seen, the performance level of the Llama3 baseline, for questions with camera perspective, is so poor that FoR identification in SG cannot help boost its performance compared to the improvements made in other language models. We should note that among our selected LLMs, Llama3 is the only one not trained with visual information; we speculate this can be a factor in LLMs’ understanding of FoR. SG is less effective in reasoning when the context and question have different perspectives despite aiding in identifying the correct FoR of the spatial description in the context. This limitation is evident in A-Split results (Table 1), where models only show significant improvement when SG aligns their preference with the question’s perspective, as seen in Qwen2-72B and GPT-4o. SG identification results are reported in the Appendix E. Still, incorporating FoR identification improves overall spatial reasoning (see the Avg column for SG+CoT in Table 3).

RQ4. How can explicit FoR identification help spatial reasoning in visualization? We evaluate SG layout diffusion to assess the impact of incorporating FoR in image generation. We focus on VISORcond metric, as it better reflects the model’s spatial understanding than the overall performance measured by VISORuncond, which is reported in Appendix C due to space limitation. Table 3 shows that adding spatial information and FoR classes (SG+GLIGEN) improves performance across all splits compared to the baseline models (GLIGEN). SG improved the model’s performance when expressions can be interpreted as relative FoR. These results align with the QA results shown in Table 1 indicating that Llama3 prefers relative FoR if dealing with the camera’s perspective. In contrast, baseline diffusion models (SD-1.5 and SD-2.1) perform better for intrinsic FoR even though GLIGEN is based on SD-2.1. This outcome might be due to GLIGEN’s reliance on bounding boxes for generating spatial configurations, which makes it struggle with intrinsic FoR due to the absence of object properties and orientation. Nevertheless, incorporating FoR information via SG-prompting improves performance across all FoR classes despite this specific bias. We provide further analysis on SG for the layout generation in Appendix D.

6 Related Work

Frame of Reference in Cognitive Studies. The concept of the frame of reference in cognitive studies was introduced by Levinson 2003 and later expanded with more diverse spatial relations (Tenbrink, 2011). Subsequent research investigated the human preferences for specific FoR classes (Edmonds-Wathen, 2012; Vukovic and Williams, 2015; Shusterman and Li, 2016; Ruotolo et al., 2016). For instance, Ruotolo et al. 2016 examined how FoR influences scene memorization and description under time constraints. Their study found that participants performed better when spatial relations were based on their position rather than external objects, highlighting a distinction between relative and intrinsic FoR.

Frame of Reference in AI. Several benchmarks have been developed to evaluate the spatial understanding of AI models in multiple modalities; for instance, textual QA (Shi et al., 2022; Mirzaee and Kordjamshidi, 2022; Rizvi et al., 2024), and text-to-image (T2I) benchmarks (Gokhale et al., 2023; Huang et al., 2023; Cho et al., 2023a, b). However, most of these benchmarks overlook the frame of reference (FoR), assuming a single FoR for all instances despite its significance in cognitive studies. Recent works in vision-language research are beginning to address this problem. For instance, Liu et al. 2023 examines FoR’s impact on visual question-answering but focuses only on limited FoR categories. Our work covers more diverse FoRs. Zhang et al. 2025 examine FoR ambiguity and understanding in vision-language models by evaluating spatial relations derived from visual input under different FoR questions. Their approach relies on images from camera perspectives, with FoR indicated in the question. In contrast, our work focuses on the reasoning of spatial relations when dealing with multiple FoRs and when there are changes in perspective in explaining the context beyond the camera’s viewpoint. Additionally, we show that explicitly identifying the FoR helps improve spatial reasoning in both question-answering and text-to-image generation, particularly when involving multiple perspectives.

7 Conclusion

Given the significance of spatial reasoning in AI applications, we introduce Frame of Reference Evaluation in Spatial Reasoning Tasks (FoREST) benchmark to evaluate Frame of Reference comprehension in textual spatial expressions via question-answering and grounding in visual modality by diffusion models. Based on this benchmark, our results reveal notable differences in FoR comprehension across LLMs and their struggle with questions that require adaptation between multiple FoRs. Moreover, the bias in FoR interpretations impacts the layout generation with LLMs for text-to-image models. To improve FoR comprehension, we propose Spatial-Guided prompting, which first generates spatial relation’s topological, distal, and directional type information in addition to FoR and includes this information in downstream task prompting. Employing SG improves the overall performance in QA tasks requiring FoR understanding and text-to-image generation.

8 Limitations

While we analyze LLMs’ shortcomings, our benchmark only highlights areas for improvement, not harming the model. The trustworthiness and reliability of the LLMs are still a research challenge. Our analysis is confined to the spatial reasoning domain and does not account for biases related to gender or race. However, we acknowledge that linguistic and cultural variations in spatial expression are not considered, as our study focuses solely on English. Extending this work to multiple languages could reveal important differences in FoR adaptation. Our analysis is still limited to the synthetic environment. Future research should consider the broader implications of the frame of reference of spatial reasoning in real-world applications. Additionally, our experiments require substantial GPU resources, limiting the selection of LLMs and constraining the feasibility of testing larger models. The computational demands also pose accessibility challenges for researchers with limited resources. We find no ethical concerns in our methodology or results, as our study does not involve human subjects or sensitive data.

References

Appendix A Dataset Statistics

The FoREST dataset statistic is provided in the Table 4.

Case A-Split A-Split for T2I FoR class C-Spilt C-split for T2I
Cow Case 792 3168 External Relative 1528 4288
Box Case 120 120 External Intrinsic 920 3680
Car Case 128 512 Internal Intrinsic 128 0
Pen Case 488 488 Internal Relative 248 0
Total 1528 4288 Total 2824 7968
Table 4: Dataset Statistic of FoREST dataset.

Appendix B Details Creation of FoREST dataset

We define the nine categories of objects selected in our dataset as indicated below in Table 5. We select sets of locatum and relatum based on the properties of each class to cover four cases of frame of reference defined in Section 3.2. Notice that we also consider the appropriateness of the container; for example, the car should not contain the bus.

Based on the selected locatum and relatum. To create an A-split spatial expression, we substitute the actual locatum and relatum objects in the Spatial Relation template. After obtaining the A-split contexts, we create their counterparts using the perspective/topology clauses to make the counterparts in C-spilt. Then, we obtain the I-A and I-C split by applying the directional template to the first occurrence of relatum when it has intrinsic directions. The directional templates are "that is facing towards," "that is facing backward," "that is facing to the left," and "that is facing to the right." All the templates are in the Table 6. We then construct the scene configuration from each modified spatial expression and send it to the simulator developed using Unity3D. Eventually, the simulator produces four visualization images for each scene configuration.

Category Object Intrinsic Direction Container
small object without intrinsic directions umbrella, bag, suitcase, fire hydrant
bog object with intrinsic directions bench, chair
big object without intrinsic direction water tank
container box, container
small animal chicken, dog, cat
big animal deer, horse, cow, sheep
small vehicle bicycle
big vehicle bus, car
tree tree
Table 5: All selected objects with two properties: intrinsic direction, affordance of being container

B.1 Simulation Details

The simulation starts with randomly placing the relatum into the scene with the orientation based on the given scene configuration. We randomly select the orientation by given scene configuration, [-40, 40] for front, [40, 140] for left, [140, 220] for back, and [220, 320] for right. Then, we create the locatum from the relatum position and move it in the spatial relation provided. If the frame of reference is relative, we move the locatum based on the camera’s orientation. Otherwise, we move it from the relatum’s orientation. Then, we check the camera’s visibility of both objects. If one of them is not visible, we repeat the process of generating the relatum until the correct placement is achieved. After getting the proper placement, we randomly choose the background from 6 backgrounds. Eventually, we repeat the procedures four times for one configuration.

B.2 Object Models and Background

For the object models and background, we find it from the unity assert store222https://assetstore.unity.com. All of them are free and available for download. All of the 3D models used are shown in Figure 4.

Refer to caption
Figure 4: All 3d models used to generate visualizations for FoREST.
{locatum} is in front of {relatum}
{locatum} is on the left of {relatum}
{locatum} is to the left of {relatum}
Spatial Relation Templates {locatum} is behind of {relatum}
{locatum} is back of {relatum}
{locatum} is on the right of {relatum}
{locatum} is to the right of {relatum}
within {relatum}
Topology Templates and inside {relatum}
and outside of {relatum}
from {relatum}’s view
relative to {relatum}
Perspective Templates from {relatum}’s perspective
from my perspective
from my point of view
relative to observer
{relatum} facing toward that camera
Orientation Templates {relatum}is facing away from the camera.
{relatum} facing left relative to the camera
{relatum} facing right relative to the camera
In the camera view, how is {locatum} positioned in relation to {relatum}?
Based on the camera perspective, where is the {locatum} from the {relatum}’s position?
Question Templates From the camera perspective, what is the relation of the{locatum} to the {relatum}?
Looking through the camera perspective, how does {locatum} appear to be oriented relative to {relatum}’s position?
Based on the camera angle, where is {locatum} located with respect to {relatum}’s location?
Table 6: All templates used to create FoREST dataset.

B.3 Textual templates

All the templates used to create FoREST are given in Table 6.

Appendix C VISOR uncond Score

VISORuncond provides the overall spatial relation score, including images with object generation errors. Since it is less focused on evaluating spatial interpretation than VISORcond, which assesses explicitly the text-to-image model’s spatial reasoning, we report VISORuncond results here in the Table 7 rather than in the main paper. The results are similar to the pattern observed in VISORuncond that the based models(SD-1.5 and SD-2.1) perform better in the relative frame of reference, while the layout-to-image models, i.e., GLIGEN, are better in the intrinsic frame of reference.

VISOR(%)
Model uncond (I) uncond (R) uncond (avg) uncond (I) uncond (R) uncond (avg)
A-Split C-Split
SD-1.5 45.4345.43 33.2233.22 43.5143.51 35.0635.06 35.6835.68 35.4035.40
SD-2.1 62.87\mathbf{62.87} 43.9043.90 59.89\mathbf{59.89} 45.98\mathbf{45.98} 46.5946.59 46.31\mathbf{46.31}
Llama3-8B + GLIGEN 46.7446.74 38.1638.16 45.3945.39 33.9833.98 39.3639.36 36.8936.89
Llama3-70B + GLIGEN 54.3354.33 46.8946.89 53.1753.17 38.0438.04 46.0446.04 42.3742.37
Llama3-8B + SG + GLIGEN (Our) 51.8351.83 43.2443.24 50.4850.48 36.2836.28 44.4344.43 40.7040.70
Llama3-70B + SG + GLIGEN (Our) 58.9258.92 47.44\mathbf{47.44} 57.1257.12 38.2338.23 48.62\mathbf{48.62} 43.8643.86
Table 7: VISORuncond score on the A-Split and C-Split where II refer to the Cow Case and Car Case where relatum has intrinsic directions, and RR refer to the Box Case and Pen case where relatum lacks intrinsic directions, avgavg is mirco-average of II and RR. cond and uncond are explained in Section 5.1.

Appendix D Analyze the improvements in SG-prompting for T2I.

Model Layout Layoutcond
Llama3-8B 85.2685.26 88.8488.84
Llama3-8B + SG 85.0485.04 88.8688.86
Llama3-70B 88.4788.47 93.1693.16
Llama3-70B + SG 91.9591.95 95.4595.45
Table 8: Layout accuracy where spatial relations are left or right relative to the camera. Layout is evaluated for all generated layouts in I-C split while Layoutcond uses the same testing examples as VISORcond.

To further explain improvements of SG-prompting in T2I task, we assess the generated bounding boxes in the I-C split for left and right relations relative to the camera since these can be evaluated using only bounding boxes without depth information. As seen is Table 8, our SG prompting improved Llama3-70B’s by 3.48%3.48\%, while Llama3-8B saw a slight decrease of 0.22%0.22\%. This evaluation was conducted on all generated layouts from the I-C split, which differs from the evaluation subset of images used for VISORcond in Table 3. We report the layoutcond score for a consistent evaluation in the same table. Layoutcond shows that Llama3-8B improves within the same evaluation subset with VISORcond. Overall, by incorporating FoR information through SG layout diffusion, Llama3 generates better spatial configurations, enhancing image generation performance.

Appendix E Frame of Reference Identification

We evaluate the LLMs’ performance in recognizing the FoR classes from given spatial expressions. The LLMs receive spatial expression, denoted as TT, and output one FoR class, FoRFoR, from the valid set of FoR classes, ForFor\in {external relative, external intrinsic, internal intrinsic, internal relative}. All in-context learning examples are in the Appendix F.

E.1 Experimental Setting

Zero-shot model. We follow the regular setting of zero-shot prompting. We only provide instruction to LLM with spatial context. The instruction prompt briefly explains each class of the FoR and candidate answers for the LLM. We called the LLM with the instruction prompt and TT to find FF.

Few-shot model. We manually craft four spatial expressions for each FoR class. To avoid creating bias, each spatial expression is ensured to fit in only one FoR class. These expressions serve as examples of our few-shotsetting. We provide these examples in addition to the instruction as a part of the prompt, followed by TT and query FF from the LLM.

Chain-of-Thought (CoT) model. To create CoT (Wei et al., 2023) examples, we modify the prompt to require reasoning before answering. Then, we manually crafted reasoning explanations with the necessary information for each example used in few-shot. Finally, we call the LLMs, adding modified instructions to updated examples, followed by TT and query FF.

Spatial-Guided Prompting (SG) model. We hypothesize that the general spatial relation types defined in Section 2 can provide meaningful information for recognizing FoR classes. For instance, a topological relation, such as “inside," is intuitively associated with an internal FoR. Therefore, we propose Spatial-Guided Prompting to direct the model in identifying the type of relations before querying FF. We revise the prompting instruction to guide the model in considering these three aspects. Then, we manually explain these three aspects. We specify the relation’s origin from the context for direction relations, such as "the left direction is relative to the observer." We hypothesize that this information helps the model distinguish between intrinsic and relative FoR. Next, we specify whether the locatum is inside or outside the relatum for topological relations. This information should help distinguish between internal and external FoR classes. Lastly, we provide the potential quantitative distance, e.g., far. This quantitative distance further encourages identifying the correct topological and directional relations. Eventually, we insert these new explanations in examples and call the model with the updated instructions followed by TT to query FF.

E.2 Evaluation Metrics

We report the accuracy of the model on the multi-class classification task. Note that the expressions in A-split can have multiple correct answers. Therefore, we consider the prediction correct when it is in one of the valid FoR classes for the given spatial expression.

Model inherently clear require template
CoT SG CoT SG
Llama3-70B 19.84 44.64 (24.80{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 24.80}) 76.72 87.39 (10.67{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 10.67})
Qwen2-72B 58.20 84.22 (26.02{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 26.02}) 88.36 93.86 (10.67{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 10.67})
GPT-4o 12.50 29.17 (16.67{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 16.67}) 87.73 90.74 (3.01{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 3.01})
Table 9: The comparison between CoT and SG prompting in C-split separated by inherently clear / required template to be clear.
A-split C-Split
Model ER-C-Split EC-Split IC-Split IR-C-Split Avg.
Gemma2-9B (0-shot) 94.1794.17 94.24\mathbf{94.24} 35.9835.98 53.9153.91 57.6657.66 60.4560.45
Gemma2-9B (4-shot) 59.5859.58 55.8955.89(38.34{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 38.34}) 72.6172.61(36.63{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 36.63}) 74.2274.22(20.31{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 20.31}) 54.4454.44(3.23{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 3.23}) 64.2964.29(3.84{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 3.84})
Gemma2-9B (CoT) 60.4960.49 60.4960.49(33.74{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 33.74}) 60.5460.54(24.57{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 24.57}) 87.5087.50(33.59{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 33.59}) 54.0354.03(3.63{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 3.63}) 65.6465.64(5.20{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 5.20})
Gemma2-9B (SG)(Our) 72.6772.67 65.8765.87(28.37{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 28.37}) 65.5465.54(29.57{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 29.57}) 53.1253.12(0.78{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 0.78}) 95.97\mathbf{95.97}(38.31{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 38.31}) 70.1370.13(9.68{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 9.68})
llama3-8B (0-shot) 60.2160.21 32.2032.20 90.1190.11 75.7875.78 0.000.00 49.5249.52
llama3-8B (4-shot) 60.1460.14 47.7747.77(15.58{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 15.58}) 54.3554.35(35.76{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 35.76}) 100.00100.00(24.22{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 24.22}) 41.1341.13(41.13{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 41.13}) 60.8160.81(11.29{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 11.29})
llama3-8B (CoT) 61.3261.32 61.0661.06(28.86{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 28.86}) 97.2897.28(7.17{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 7.17}) 100.00100.00(24.22{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 24.22}) 36.2936.29(36.29{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 36.29}) 73.6673.66(24.14{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 24.14})
llama3-8B (SG) (Our) 62.9562.95 63.2963.29(31.09{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 31.09}) 94.5794.57(4.46{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 4.46}) 100.00100.00(24.22{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 24.22}) 43.5543.55(43.55{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 43.55}) 75.3575.35(25.83{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 25.83})
llama3-70B (0-shot) 84.2384.23 74.0874.08 9.579.57 92.1992.19 68.5568.55 61.1061.10
llama3-70B (4-shot) 78.4778.47 81.8181.81(7.72{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 7.72}) 64.8964.89(55.33{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 55.33}) 100.00100.00(7.81{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 7.81}) 75.8175.81(7.26{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 7.26}) 80.6380.63(19.53{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 19.53})
llama3-70B (CoT) 69.1169.11 72.0572.05(2.03{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 2.03}) 97.0797.07(87.50{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 87.50}) 100.00100.00(7.81{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 7.81}) 79.4479.44(10.89{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 10.89}) 87.1487.14(26.04{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 26.04})
llama3-70B (SG) (Our) 76.5076.50 78.2178.21(4.12{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 4.12}) 97.6197.61(88.04{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 88.04}) 100.00100.00(7.81{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 7.81}) 72.1872.18(3.63{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 3.63}) 87.0087.00(25.90{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 25.90})
Qwen2-7B (0-shot) 83.6483.64 79.9779.97 59.2459.24 77.3477.34 40.7340.73 64.3264.32
Qwen2-7B (4-shot) 61.1261.12 50.5250.52(29.45{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 29.45}) 65.7665.76(6.52{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 6.52}) 93.7593.75(16.41{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 16.41}) 56.0556.05(15.32{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 15.32}) 66.5266.52(2.20{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 2.20})
Qwen2-7B (CoT) 72.1272.12 70.8170.81(9.16{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 9.16}) 63.8063.80(4.57{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 4.57}) 99.2299.22(21.88{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 21.88}) 51.6151.61(10.89{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 10.89}) 71.3671.36(7.04{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 7.04})
Qwen2-7B (SG) 70.6170.61 68.0068.00(11.98{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 11.98}) 71.2071.20(11.96{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 11.96}) 88.2888.28(10.94{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 10.94}) 57.2657.26(16.53{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 16.53}) 71.1871.18(6.86{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 6.86})
Qwen2-72B (0-shot) 64.4664.46 62.7062.70 100.00100.00 100.00100.00 39.1139.11 75.4575.45
Qwen2-72B (4-shot) 79.1279.12 78.7378.73(16.03{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 16.03}) 99.3599.35(0.65{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 0.65}) 87.5087.50(12.50{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 12.50}) 87.1087.10(47.98{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 47.98}) 88.1788.17(12.72{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 12.72})
Qwen2-72B (CoT) 88.5488.54 88.8788.87(26.18{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 26.18}) 89.5789.57(10.43{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 10.43}) 93.7593.75(6.25{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 6.25}) 83.4783.47(44.35{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 44.35}) 88.9188.91(13.46{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 13.46})
Qwen2-72B (SG) 90.5190.51 90.1890.18(27.49{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 27.49}) 93.2693.26(6.74{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 6.74}) 98.4498.44(1.56{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 1.56}) 85.0885.08(45.97{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 45.97}) 91.7491.74(16.29{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 16.29})
GPT3.5 (0-shot) 83.1183.11 88.1588.15 17.5017.50 70.3170.31 41.1341.13 54.2754.27
GPT3.5 (4-shot) 61.2561.25 48.9548.95(39.20{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 39.20}) 62.7262.72(45.22{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 45.22}) 100.00100.00(29.69{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 29.69}) 28.6328.63(12.50{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 12.50}) 60.0760.07(5.80{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 5.80})
GPT3.5 (CoT) 66.5566.55 66.6266.62(21.53{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 21.53}) 96.8596.85(79.35{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 79.35}) 100.00100.00(29.69{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 29.69}) 50.8150.81(9.68{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 9.68}) 78.5778.57(24.30{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 24.30})
GPT3.5 (SG) (Our) 70.6170.61 73.3073.30(14.86{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 14.86}) 92.9392.93(75.43{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 75.43}) 99.2299.22(28.91{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 28.91}) 49.1949.19(8.06{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 8.06}) 78.6678.66(24.39{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 24.39})
GPT4o (0-shot) 73.8273.82 71.2771.27 98.8098.80 100.00100.00 70.5670.56 85.1685.16
GPT4o (4-shot) 66.2366.23 67.8767.87(3.40{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 3.40}) 98.7098.70(0.11{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 0.11}) 100.00100.00(0.00{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 0.00}) 78.6378.63(8.06{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 8.06}) 86.3086.30(1.14{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 1.14})
GPT4o (CoT) 72.4472.44 72.7772.77(1.51{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 1.51}) 100.00100.00(1.20{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 1.20}) 100.00100.00(0.00{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 0.00}) 73.7973.79(3.23{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 3.23}) 86.6486.64(1.48{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 1.48})
GPT4o (SG) (Our) 76.4476.44 74.6774.67(3.40{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 3.40}) 97.7297.72(1.09{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 1.09}) 100.00100.00(0.00{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 0.00}) 68.5568.55(2.02{\color[rgb]{1,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{1,0,0}\downarrow 2.02}) 85.2385.23(0.08{\color[rgb]{0,1,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,1,0}\uparrow 0.08})
Table 10: Accuracy results report from FoR Identification with LLMs. The correct prediction is one of the valid FoR classes for the given spatial expression. All FoR classes are external relative (ER), external intrinsic (EI), internal intrinsic (II), and internal relative (IR).
Refer to caption
(a) Results of Cow Case in A-Split.
Refer to caption
(b) Results of Car Case in A-Split.
Refer to caption
(c) Results of Box Case in A-Split.
Refer to caption
(d) Results of Pen Case in A-Split.
Figure 5: Red shows the wrong FoR identifications, and green shows the correct ones. The dark color is for relative FoRs, while the light color is for intrinsic FoRs. The round shape is for the external FoRs, while the square is for internal FoRs. The depth of the plots shows the four FoRs, i.e., external relative, external intrinsic, internal intrinsic, and internal relative, from front to back.
Refer to caption
(a) Results of Cow Case in A-Split.
Refer to caption
(b) Results of Car Case in A-Split.
Refer to caption
(c) Results of Box Case in A-Split.
Refer to caption
(d) Results of Pen Case in A-Split.
Figure 6: Red shows the wrong FoR identifications, and green shows the correct ones. The dark color is for relative FoRs, while the light color is for intrinsic FoRs. The round shape is for the external FoRs, while the square is for internal FoRs. The depth of the plots shows the four FoRs, i.e., external relative, external intrinsic, internal intrinsic, and internal relative, from front to back. This plot is the result of the rest of LLMs.

E.3 Results

E.3.1 FoR Inherently Bias in LLMs

C-spilt. The zero-shot setting reflects the LLMs’ inherent bias in identifying FoR. Table 10 presents the accuracy for each FoR class in C-split, where sentences explicitly include information about topology and perspectives. We found that some models strongly prefer specific FoR classes. Notably, Gemme2-9B achieves a near-perfect accuracy on external relative FoR but performs poorly on other classes, especially external intrinsic, indicating a notable bias towards external relative. In contrast, GPT4o and Qwen2-72B perform exceptionally in both intrinsic FoR classes. However, they perform poorly in the relative FoRs.

A-spilt. We examine the FoR bias in the A-split. Based on the results in Table 10, we plotted the top-3 models’ results (Gemma2-9B, Llama3-70B, and GPT4o) for a more precise analysis in Figures 5. The plots show the frequencies of each FoR category. According to the plot, Gemma and GPT have strong biases toward external relative and external intrinsic, respectively. This bias helps Gemma2 perform well in the A-split since all spatial expressions can be interpreted as external relative. However, GPT4o’s bias leads to errors when intrinsic FoRs aren’t valid, as in the Box and Pen cases (see plots (c) and (d)). Llama3 exhibits different behavior, showing a bias based on the relatum’s properties, specifically the relatum’s affordance as a container. In cases where relatum cannot serve as containers, i.e., Cow and Pen cases, Llama3 favors external relative. Conversely, Llama3 tends to favor external intrinsic when the relatum has the potential to be a container.

E.3.2 Behavior with ICL variations

C-spilt. We evaluate the models’ behavior under various in-context learning (ICL) methods. As observed in Table 10, the few-shot method improves the performance of the zero-shot method across multiple LLMs by reducing their original bias toward specific classes. Reducing the bias, however, lowers the performance in some cases, such as the performance of Gemma 2 in ER class. One noteworthy observation is that while the CoT prompting generally improves performance in larger LLMs, it is counterproductive in smaller models for some FoR classes. This suggests that the smaller models have difficulty inferring FoR from the longer context. This negative effect also appears in SG prompting, which uses longer explanations. Despite performance degradation in particular classes of small models, SG prompting performs exceptionally well across various models and achieves outstanding performance with Qwen2-72B. We further investigate the performance of CoT and SG prompting. As shown in Table 9, CoT exhibits a substantial difference in performance between contexts with inherently clear FoR and contexts requiring the template to clarify FoR ambiguity. This implies that CoT heavily relies on the specific template to identify FoR classes. In contrast, SG prompting demonstrates a smaller gap between these two scenarios and significantly enhances performance over CoT in inherently clear FoR contexts. Therefore, guiding the model to provide characteristics regarding topological, distance, and directional types of relations improves FoR comprehension.

A-spilt. We use the same Figure 5 to observe the behavior when applying ICL. The A-split shows minimal improvement with ICL variations, though some notable changes are observed. With few-shot, all models show a strong bias toward external intrinsic FoR, even when the relatum lacks intrinsic directions, i.e., Box and Pen cases. This bias appears even in Gemma2-9B, which usually behaves differently. This suggests that the models pick up biases from the examples despite efforts to avoid such patterns. However, CoT reduces some bias, leading LLMs to revisit relative, which is generally valid across scenarios. In Gemma2, the model predicts relative FoR where the relatum has intrinsic directions, i.e., Cow and Car cases. Llama3 behaves similarly in cases where the relatum cannot act as a container, i.e., Cow and Pen cases. GPT4o, however, does not depend on the relatum’s properties and shows slight improvements across all cases. Unlike CoT, our SG prompting is effective in all scenarios. It significantly reduces biases while following a similar pattern to CoT. Specifically, SG prompting increases external relative predictions for Car and Cow in Gemma2-9B, and for Cow and Pen in Llama3-70B. Nevertheless, GPT4o shows only a slight bias reduction. However, Our proposed method improves the overall performance of most models, as shown in Table 10. The Llama3-70B behaviors are also seen in LLama3-8B and GPT3.5. The plots for LLama3-8B and GPT3.5 are in Figure 6.

E.3.3 Experiment with different temperatures

We conducted additional experiments to further investigate the impact of temperature on the biased interpretation of the model in the A-split of our dataset. As presented in Table 11, comparing distinct temperatures (0 and 1) revealed a shift in the distribution. The frequencies of the classes experienced a change of up to 10%. However, the magnitude of this change is relatively minor, and the relative preferences for most categories remained unchanged. Specifically, the model exhibited the highest frequency responses for the cow, car, and pen cases, even with higher frequencies in certain settings. Consequently, a high temperature does not substantially alter the diversity of LLMs’ responses to this task, which is an intriguing finding.

Model ER EI II IR
temp-0 temp-1 temp-0 temp-1 temp-0 temp-1 temp-0 temp-1
Cow Case
0-shot 75.38 87.12 23.86 12.50 0.76 0.13 0.00 0.25
4-shot 0.00 15.66 100.00 84.34 0.00 0.00 0.00 0.00
CoT 31.82 49.87 68.18 49.87 0.00 0.13 0.00 0.13
SG 51.39 70.45 48.61 29.42 0.00 0.00 0.00 0.13
Box Case
0-shot 22.50 41.67 77.50 58.33 0.00 0.13 0.00 0.25
4-shot 0.00 0.00 100.00 100.00 0.00 0.00 0.00 0.00
CoT 0.00 5.83 100.00 94.17 0.00 0.00 0.00 0.00
SG 11.67 33.33 88.33 66.67 0.00 0.00 0.00 0.00
Car Case
0-shot 55.20 68.24 49.01 31.15 0.79 0.61 0.00 0.00
4-shot 0.60 5.94 99.40 94.06 0.00 0.00 0.00 0.00
CoT 19.64 38.52 80.16 61.27 0.20 0.20 0.00 0.00
SG 44.25 56.97 55.75 43.03 0.00 0.00 0.00 0.00
Pen Case
0-shot 90.62 96.88 9.38 3.12 0.00 0.61 0.00 0.00
4-shot 0.00 7.03 100.00 92.97 0.00 0.00 0.00 0.00
CoT 17.19 28.91 82.81 71.09 0.20 0.20 0.00 0.00
SG 48.31 57.81 54.69 42.19 0.00 0.00 0.00 0.00
Table 11: The results between two different temperatures of Llam3-70B on the A-spilt of FoREST. The number shows the percentage frequency of responses from the model.

Appendix F In-context learning

F.1 FoR Identification

We provide the prompting for each in-context learning. The prompting for zero-shot and few-shot is provided in Listing 1. The instruction answer for these two in-context learning is “Answer only the category without any explanation. The answer should be in the form of {Answer: Category.}"

For the Chain of Thought (CoT), we only modified the instruction answer to “Answer only the category with an explanation. The answer should be in the form of {Explanation: Explanation Answer: Category.}" Similarly to CoT, we only modified the instruction answer to “Answer only the category with an explanation regarding topological, distance, and direction aspects. The answer should be in the form of {Explanation: Explanation Answer: Category.}", respectively. The example responses are provided in Listing 4 for Spatial Guided prompting.

# Instruction to find frame of reference class of given context
"""
Instruction:
You specialize in language and spatial relations, specifically in the frame of context (multiple perspectives in the spatial relation). Identify the frame of reference category given the following context. There are four classes of the frame of reference (external intrinsic, internal intrinsic, external relative, internal relative). Note that the intrinsic direction refers to whether the model has the front/back by itself. (Example: a bird, human. Counter Example: a ball, a box). "
External intrinsic. The spatial description of an entity A relative to another entity B, where (1) A is not contained by B, (2) the spatial relation is based on B’s facing orientation (intrinsic direction) if B has one.
Internal intrinsic. The spatial description of an entity A relative to another entity B, where (1) A is contained by B, (2) the spatial relation is based on B’s facing orientation (intrinsic direction) if B has one.
External relative. The spatial description of an entity A relative to another entity B, where (1) A is not contained by B, (2) The spatial relation is relative to entity C, where C is not involved in the spatial relation and only provides the perspective. An example of C is the camera.
Internal relative. The spatial description of an entity A relative to another entity B, where (1) A is contained by B, (2) The spatial relation is relative to entity C, where C is not involved in the spatial relation and only provides the perspective. An example of C is the camera.
{Instruction answer}
# Normal Instruction answer: Find the frame of reference category without any explanations. The answer should be in the form of Answer: Category.
# COT Instruction answer: Find the frame of reference category with an explanation. The answer should be in the form of Explanation: Explanation. Answer: Category.
# SG Instruction answer: Find the frame of reference category with an explanation. Additionally, include three more explanations relating to topology, distance, and direction from the given context. The answer should be in the form of \"Explanation: Explanation. Topology: Explanation. Distance: Explanation. Direction: Explanation. Answer: Category.
Context: {spatial exprssion}
""""
Listing 1: Prompt for finding the frame of reference class of given context.
## These are four examples used in few-shot, CoT, and SG
# Example 1: The bird is outside the car and in front of the car relative to the car.
# Few-shot response: Answer: external intrinsic
# CoT response: Explanation: The car does not contain the bird, and the spatial relation is relative to the car’s facing orientation. The intrinsic direction of the car is relevant in this context, as the description \"in front of\" implies a direction relative to the sheep’s front or back. Answer: external intrinsic.
# SG response: Explanation: The car does not contain the bird, and the spatial relation is relative to the car’s facing orientation. The intrinsic direction of the car is relevant in this context, as the description \"in front of\" implies a direction relative to the sheep’s front or back. Topology: The bird is outside the car. Distance: There is little distance between the bird and the car. Direction: The bird is in the front direction of the car (entity B). Answer: external intrinsic.
# Example 2: The bird is inside of the car and left of the car from the car’s perspective.
# Few-shot response: Answer: internal intrinsic
# CoT response: Explanation: In this scenario, the bird is contained within the car, meaning it is an internal relationship. Additionally, the description specifies the bird’s position relative to the car, indicating that the spatial relation is relative to the car itself, considering its intrinsic direction. Answer: internal intrinsic.
# SG response: Explanation: In this scenario, the bird is contained within the car, meaning it is an internal relationship. Additionally, the description specifies the bird’s position relative to the car, indicating that the spatial relation is relative to the car itself, considering its intrinsic direction. Topological: The bird is inside the car. Distance: The bird is in the same location as the car. Direction: The bird is in the left direction of the car (entity B). Answer: internal intrinsic.
# Example 3: The box is inside and at the back of the room.
# Few-shot response: Answer: internal relative
# CoT response: Explanation: In this context, the box (entity A) is described as inside the room (entity B), fulfilling the first condition for an internal frame of reference. However, the spatial relation given (\" at the back\") cannot originate from the room since the room does not have intrinsic direction, implying it is relative to an external observer or reference point (C), such as a camera or a person viewing the scene. Answer: internal relative
# SG response: Explanation: In this context, the box (entity A) is described as inside the room (entity B), fulfilling the first condition for an internal frame of reference. However, the spatial relation given (\" at the back\") cannot originate from the room since the room does not have intrinsic direction, implying it is relative to an external observer or reference point (C), such as a camera or a person viewing the scene. Topological: The box is inside the room. Distance: The box is in the same location as the room. Direction: From the observer’s perspective, the box is located at the back of the room. Answer: internal relative
# Example 4: A phone is on the left of a tablet from my perspective.
# Few-shot response: Answer: external relative
# CoT response: Explanation: In this context, the spatial relation is described from the perspective of an observer (C) who is not involved in the spatial relation. The phone (A) is not contained by the tablet (B), and the spatial relation \"left of\" is relative to the observer’s perspective, not the tablet’s facing orientation. Answer: external relative.
# SG response: Explanation: In this context, the spatial relation is described from the perspective of an observer (C) who is not involved in the spatial relation. The phone (A) is not contained by the tablet (B), and the spatial relation \"left of\" is relative to the observer’s perspective, not the tablet’s facing orientation. Topological: The phone is not contained by the tablet. Distance: There is some distance between the phone and the tablet. Direction: From the observer’s perspective, the phone is located to the left of the tablet. Answer: external relative.
Listing 2: Spatial expression examples using for few-shots, Chain-of-Thought (CoT), and Spatial Guide (SG) prompting

F.2 Question Answering Example

’’
#Context: The bird is outside the car and in front of the car relative to the car. The car is facing toward the camera.
#Question: Based on the camera’s perspective, where is the bird from the car position in the scene?
#SG information (used in SG + CoT): Frame of reference Explanation: The car does not contain the bird, and the spatial relation is relative to the car’s facing orientation. The intrinsic direction of the car is relevant in this context, as the description \"in front of\" implies a direction relative to the sheep’s front or back. Topology: The bird is outside the car. Distance: There is little distance between the bird and the car. Direction: The bird is in the front direction of the car (entity B). Frame of reference: external intrinsic.
#Normal response: Answer: front
#CoT / (SG + COT) response:Explanation: Based on the context, the bird’s position is in the front direction of the car. The car is facing toward the camera. Then, the car’s front direction is the camera’s front direction. Therefore, the bird’s position is in front of the car’s position from the camera’s perspective. Answer: front
#Context: The bird is inside the car and left of the car from the car’s perspective. The car is facing to the right relative to the camera. #Question: Based on the camera’s perspective, where is the bird from the car’s position?
#SG information (used in SG + CoT): Frame of reference Explanation: In this scenario, the bird is contained within the car, meaning it is an internal relationship. Additionally, the description specifies the bird’s position relative to the car, indicating that the spatial relation is relative to the car itself, considering its intrinsic direction. Topological: The bird is inside the car. Distance: The bird is in the same location as the car. Direction: The bird is in the left direction of the car (entity B). Frame of reference: internal intrinsic.
#Normal response: Answer: back
#CoT / (SG + COT) response: Explanation: Based on the context, the bird’s position is in the left direction of the car. The car is facing to the right. Then, the car’s left direction is the camera’s back direction. Therefore, the bird’s position is to the back of the car’s position from the camera’s perspective. Answer: back
Context: The box is inside and to the right of the room from the observer’s perspective.
Question: From the observer’s perspective, what is the spatial relation of the box to the room?
#SG information (used in SG + CoT): Frame of reference Explanation: In this context, the box (entity A) is described as inside the room (entity B), fulfilling the first condition for an internal frame of reference. However, the spatial relation given (\" at the back\") cannot originate from the room since the room does not have intrinsic direction, implying it is relative to an external observer or reference point (C), such as a camera or a person viewing the scene. Topological: The box is inside the room. Distance: The box is in the same location as the room. Direction: From the observer’s perspective, the box is located at the back of the room. Frame of reference: internal relative.
#Normal response: Answer: right
#CoT / (SG + COT) response: Explanation: Based on the context, the box is to the right of the room from the camera’s direction. Therefore, the box’s position is to the right of the room’s position from the observer’s perspective. Answer: right
Context: A phone is to the left of a tablet from my perspective. The tablet is facing to the right. Question: From my perspective, what is the spatial relation of the phone to the tablet?
#SG information (used in SG + CoT): Frame of Reference Explanation: In this context, the spatial relation is described from the perspective of an observer (C) who is not involved in the spatial relation. The phone (A) is not contained by the tablet (B), and the spatial relation \"left of\" is relative to the observer’s perspective, not the tablet’s facing orientation. Topological: The phone is not contained by the tablet. Distance: There is some distance between the phone and the tablet. Direction: From the observer’s perspective, the phone is located to the left of the tablet. Frame of Reference: external relative.
#Normal response: Answer: left
#CoT / (SG + COT) response: Explanation: Based on the context, the phone is to the left of the tablet from my perspective. The direction of the tablet is not relevant in this context since the left relation is from my perspective. Therefore, from my perspective, the phone is to the left of the tablet. Answer: left
’’’’
Listing 3: Spatial expression examples using for few-shots, Chain-of-Thought (CoT), and Spatial Guide (SG) prompting for question-answering.

F.3 Text to Layout

# Instruction for generating bounding box
"""
Your task is to generate the bounding boxes of objects mentioned in the caption.
The image is size 512x512. The bounding box should be in the format of (x, y, width, height). Please considering the frame of reference of caption and direction of reference object if possible. If needed, you can make the reasonable guess.
"""
Listing 4: Prompt for generating bounding coordinates to use as the layout for layout-to-image models.