This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\commission

3, 1 \workinggroupXX/YY \icwg

Tree-GPT: Modular Large Language Model Expert System for Forest Remote Sensing Image Understanding and Interactive Analysis

S. Q. Dua,b,c    S.J. Tanga,b,c Corresponding author    W. X. Wang a,b,c    X. M. Lia,b,c    R. Z. Guoa,b,c a Research Institute for Smart Cities, School of Architecture and Urban Planning, Shenzhen University, Shenzhen, P.R. China
b State Key Laboratory of Subtropical Building and Urban Science, Shenzhen, P.R. China
c Guangdong–Hong Kong-Macau Joint Laboratory for Smart Cities, P.R. China
Abstract

This paper introduces a novel framework, Tree-GPT, which incorporates Large Language Models (LLMs) into the forestry remote sensing data workflow, thereby enhancing the efficiency of data analysis. Currently, LLMs are unable to extract or comprehend information from images and may generate inaccurate text due to a lack of domain knowledge, limiting their use in forestry data analysis. To address this issue, we propose a modular LLM expert system, Tree-GPT, that integrates image understanding modules, domain knowledge bases, and toolchains. This empowers LLMs with the ability to comprehend images, acquire accurate knowledge, generate code, and perform data analysis in a local environment. Specifically, the image understanding module extracts structured information from forest remote sensing images by utilizing automatic or interactive generation of prompts to guide the Segment Anything Model (SAM) in generating and selecting optimal tree segmentation results. The system then calculates tree structural parameters based on these results and stores them in a database. Upon receiving a specific natural language instruction, the LLM generates code based on a thought chain to accomplish the analysis task. The code is then executed by an LLM agent in a local environment and . For ecological parameter calculations, the system retrieves the corresponding knowledge from the knowledge base and inputs it into the LLM to guide the generation of accurate code. We tested this system on several tasks, including Search, Visualization, and Machine Learning Analysis. The prototype system performed well, demonstrating the potential for dynamic usage of LLMs in forestry research and environmental sciences.

keywords:
Remote Sensing, Deep Learning, Forestry, Large Language Model, Individual Tree Segmentation, Tree Factor Estimation, Segment Anything Model.

1 Introduction

Analysis of forest remote sensing data is essential for various applications in forestry and environmental sciences[Turner et al., 2006, Suratno et al., 2009, Indirabai et al., 2019, Sun et al., 2019]. The analysis of forest remote sensing data is often subject to temporal constraints, necessitating the exploration of more efficient method. Recent advancements in artificial intelligence have given rise to Large Language Models (LLMs), which are machine learning algorithms capable of comprehending human natural language instruction and generating coherent text [Zhao et al., 2023]. By utilizing natural language as an interface, researchers can employ LLMs to generate data analysis code in a matter of minutes, a task that previously required several tens of minutes to complete. Consequently, the integration of LLMs into the analysis of forest remote sensing data is being considered as a means of enhancing efficiency in a similar manner.

LMMs like GPT series have left a profound impression due to their powerful intelligence [Ouyang et al., 2022]. Naturally, we are curious about whether we can employ similar models to perform intelligent analysis of forest remote sensing data. However, the limited ability of LLMs to comprehend images and generate accurate domain-specific knowledge hinders their application in forest remote sensing data analysis. Vision serves as the primary source of information for human cognition. However, the majority of language models struggle to comprehend visual information, limiting their applicability. In recent years, researchers have investigated methods to enable LLMs to process multimodal information. Among these studies, GPT-4 has emerged as the leading model, exhibiting remarkable performance in understanding and accurately describing natural images [OpenAI, 2023]. Additionally, open-source models such as MiniGPT4 have proposed their own solutions [Zhu et al., 2023]. However, these approaches exhibit suboptimal performance on remote sensing imagery due to the fact that the majority of these models are trained on natural images. To address this limitation, researchers have proposed multimodal remote sensing datasets and trained models that capable of understanding remote sensing images and performing natural language-based question answering [Lobry et al., 2020, Chappuis et al., 2022]. These models can understand natural language and provide answers regarding the presence of specific concepts in the images. However, due to their complexity, these model are still not capable of precise quantitative analysis or executing data analysis tasks through code generation. Furthermore, most language models exhibit limited performance when generating text in specific professional domains due to a lack of fine-tuning for those domains. A phenomenon known as “hallucination” is also commonly observed in language models, where they generate information that does not exist in reality, such as providing incorrect interpretations of natural laws [Ji et al., 2023]. Consequently, the lack of remote sensing image understanding ability and accurate professional knowledge are significant issues that restrict the application of general language models in specialized domains.

To address the first issue and enable LLMs to understand forest remote sensing images, we are considering the integration of a separate image understanding module. This module is built to convert the implicit information contained in the image into explicit, structured information that LLMs can understand. For forest remote sensing data analysis, it typically involves two primary tasks: the individual tree crown (ITC) segmentation and calculation of tree structural parameters. Among them, the ITC segmentation remains a challenging problem. Current research has explored various methods based on point clouds or images for ITC segmentation. Among them, image-based segmentation methods tend to employ supervised learning algorithms like Mask R-CNN for accurate instance segmentation[Beloiu et al., 2023]. These methods require a large number of high-precision training samples. To reduce the cost of our approach, we are considering whether unsupervised or interactive methods can be used for ITC segmentation to generate structured information that LLMs can comprehend.

Recently, image representation learning [Vaswani et al., 2017, Dosovitskiy et al., 2020] and self-supervised learning [He et al., 2022] have advanced significantly, along with the availability of large-scale data and computational resources [Kirillov et al., 2023]. These developments have enabled the learning of universal features that can be transferred across different domains from massive image collections. In this context, a Large Vision Model named Segment Anything Model (SAM) was proposed [Kirillov et al., 2023], which can perform zero-shot image segmentation guided by prompts. However, SAM still faces two challenges in individual tree segmentation task: 1) the segmentation quality of SAM is highly sensitive to the input prompt, and the segmentation granularity is not controllable without prior knowledge. 2) SAM generates many redundant masks in its segmentation output, which hinders the identification of individual trees. This paper aims to overcome these challenges and leverage the potential of SAM for single-tree segmentation tasks by proposing a framework that does not require retraining and can be easily adapted to new scenarios.

To address the second issue of empowering LLMs with domain-specific knowledge and task execution capabilities, we are considering building a specialized knowledge base and an execution agent for LLMs. The specialized knowledge base is a widely used solution to tackle the issue of hallucination in LLMs. Meanwhile, the execution agent acts as a framework that connects LLMs with the local execution environment. LLMs can utilize the execution agent to execute generated code in the local development environment, perform data manipulation, and generate visualization or data analysis results. Currently, there are existing frameworks for knowledge base construction and task execution for LLMs. However, designing the optimal framework is still an open question. For the specific task of forest remote sensing data analysis, it is worth discussing how to adapt the existing frameworks, such as Prompt Engineering, to suit the characteristics of this particular task.

Following above ideas, we propose a novel framework, Tree-GPT. The purpose of Tree-GPT is to build an expert system capable of understanding forest remote sensing images, possessing domain-specific knowledge in forest ecology, and generating code based on prompts to automatically execute data analysis tasks. The core of this system consists of an LLM as the reasoning engine, an image understanding module to convert image information into text, a Domain Knowledge Base to store professional knowledge for LLM retrieval, and an LLM execution agent for code execution. More specifically:

  • LLM: The LLM and its accompanying backend execution framework are pivotal components of Tree-GPT. In terms of model selection, the study utilizes the cloud-deployed OpenAI GPT-4 model, known for its high performance. However, in theory, an LLM deployed locally would also be suitable.

  • Image Understanding: As the second key module of Tree-GPT, the image understanding module encompasses two functions: tree segmentation and calculation of tree structural parameters. To enable low-cost algorithm transfer, the study employs the Structure-Aware Masking (SAM) as the core and designs a new tree segmentation method to address the challenges associated with SAM. This method starts by employing a pre-trained tree detection model to locate and outline trees (or uses an interactive approach) to generate an initial prompt. Subsequently, the study inputs the initial prompt and a grid prompt into SAM, resulting in redundant tree segmentation outcomes. To refine the segmentation results, a bipartite graph matching model is utilized, leveraging the initial prompt as a guide. After obtaining the segmentation results, additional remote sensing data is incorporated to compute tree structural parameters, which are stored in a relational database for subsequent data analysis access.

  • Domain Knowledge Base: As the third key module of Tree-GPT, the Domain Knowledge Base is a vector database that stores forest ecology domain-specific knowledge in embedded text form. This knowledge base allows for retrieval of crucial information through keyword searches and utilizes it as prompts to guide the LLM in generating accurate professional knowledge outputs.

  • LLM Execution Agent: Lastly, as the execution module of Tree-GPT, the LLM execution agent serves as a backend framework that connects the LLM’s code outputs with the local runtime environment. The study has tailored the Prompt Engineering framework to address the specific tasks of forest remote sensing data analysis, including information retrieval, result visualization, and data analysis. Task-oriented prompt design ensures the generation of precise code, facilitating efficient and effective data analysis.

Refer to caption
Figure 1: Overall Framework of the Tree-GPT.

In conclusion, Tree-GPT offers a user-friendly approach that enables real-time data updates and analysis. This reduces the time required for the processing workflow and allows researchers to allocate more resources towards data interpretation and the understanding of ecological patterns. With its understanding and generation capabilities provided by Large Language Models, Tree-GPT offers a pertinent contribution to the field of tree ecological parameter extraction.

2 Related Works

2.1 Individual Tree Crown (ITC) Segmentation

Efficient and automated ITC segmentation methods are crucial for calculating tree factors [Yang et al., 2019]. Existing methods for ITC segmentation can be classified based on the type of data used, including point cloud based methods and image based method. Current UAV point cloud based methods often utilize unsupervised clustering techniques. Since the highest point of a tree (which can be regarded as the tree’s center) is easily identifiable in the point cloud, these methods typically determine the tree center by leveraging the elevation differences. They then perform point cloud clustering using the center as a seed point to segment the tree contour [Strîmbu and Strîmbu, 2015]. However, accurately describing tree boundaries becomes challenging due to the lack of texture information in point clouds. As a result, most methods rely on strong prior assumptions such as the conical shape assumption to differentiate between points at the boundaries of trees [Strîmbu and Strîmbu, 2015, Yang et al., 2020, Qian et al., 2023]. These methods generally achieve acceptable results in most scenarios. However, in subtropical or tropical regions, trees tend to have broader and more expansive shapes with large crown areas [Lutz et al., 2012]. As a result, many trees’ growth patterns deviate from the aforementioned prior assumptions, which limits the generalization capability of these methods when applied to new scenarios, necessitating repeated parameter tuning.

Based on image data, tree segmentation methods primarily utilize the color and texture differences of different trees in RGB images or the tree height differences in the canopy height model (CHM) to segment individual trees. Some studies have employed traditional morphological image segmentation methods, such as watershed algorithms and their variants, to achieve tree segmentation in RGB or CHM images [Tochon et al., 2015, Wagner et al., 2018, Yang et al., 2019]. However, the limited utilization of morphological image features hinders the accuracy of these segmentation methods. In recent years, research has focused on deep learning-based single-tree segmentation methods due to the advancement of deep learning techniques. Compared to traditional image segmentation methods, deep learning networks possess stronger feature extraction capabilities and can learn features with certain generalization properties, thus enhancing the generalization of the methods [LeCun et al., 2015]. Currently, most deep learning-based single-tree segmentation methods are based on the Mask-RCNN network [He et al., 2017, Beloiu et al., 2023]. Some methods have improved edge extraction by enhancing the loss function [Zhang et al., 2022], while others have considered incorporating multi-modal feature extraction techniques within the Mask-RCNN framework. They simultaneously use texture, color, and elevation information from both RGB and CHM images to improve tree segmentation accuracy [Li et al., 2022]. Although deep learning-based instance segmentation methods for single-tree segmentation show significant improvements over traditional image segmentation methods in terms of robustness and generalization, most supervised deep learning methods suffer from a fundamental limitation: their performance is heavily constrained by the training data. Since deep learning is essentially about finding a set of factors that best fit the distribution of the training samples, the performance of the network is determined by the extent to which the probability distribution of the training samples aligns with the real-world distribution [Goodfellow et al., 2016]. For single-tree segmentation tasks, high-precision open-source tree contour datasets are severely lacking. Moreover, distinguishing and annotating tree contours is extremely costly compared to natural images, making it difficult to rapidly annotate a large amount of data to train a tree segmentation network with strong generalization and zero-shot transfer capability.

3 Method

Tree-GPT is a specialized system designed to process remote sensing data from forests. It comprises a GPT-4-based LLM, an image understanding module, a Domain Knowledge Base, and an LLM execution agent. The Tree-GPT workflow operates as follows: given a natural language instruction LL and a remote sensing image II of the study area, the image understanding module processes II and collaborates with other remote sensing data from the study area to generate tree structure parameters that can be used as inputs for the LLM and for data analysis. These parameters are stored in a relational database RR. Simultaneously, the professional domain knowledge base module assesses LL to determine if it is a knowledge-based question and whether to invoke the professional knowledge base module to answer it. If LL is not a knowledge-based question, it proceeds to the LLM execution agent module.

In the LLM execution module, if LL does not involve code execution tasks, the output result of the LLM is directly returned in the FastChat frontend. If the result contains code execution tasks, the corresponding code is generated based on prompt templates and executed using the LangChain agent module in a local runtime environment. During execution, data visualization or analysis results are generated by calling image II and database RR.

Refer to caption
Figure 2: Visualization of individual tree crown(ITC) segmentation results. The first line shows tree crown segmentation results (segmentation in 2D). The second line shows individual tree point cloud segmentation results (3D segmentation by projecting 2D results to point cloud).

3.1 Image Understanding Module

Mask Generation:The Tree-GPT model utilizes a tree target detection framework and a bipartite graph matching method to control the segmentation granularity of SAM. This generates tree segmentation results that correspond one-to-one with the tree position. Specifically, the single-tree segmentation module uses points and target frames as SAM’s Prompt at the same time to guide the generation of tree Masks. Among them, the target detection module can use automatic or manual methods to generate tree target detection frames, which are used to guide SAM for tree segmentation and assign semantic information to the segmentation results. The automatic target frame generation method uses RetinaNet [Lin et al., 2017] as the target detection framework with ResNet50 [He et al., 2016] as the backbone network and DeepForest [Weinstein et al., 2019] pretrained weights. Manual annotation of tree target frames can be used as a supplement to automatic tree target detection to improve the accuracy of tree detection in an interactive manner. In addition to the target frame, we use a 48 ×\times 48 grid sampling point as the basic Prompt for SAM to deal with situations where BBox Prompt guidance fails. This generates enough Masks for subsequent matching with the target frame and obtains optimal segmentation results of the tree. After obtaining the Prompt, we input Point and BBox Prompt into SAM respectively and merge the two sets of tree segmentation results to obtain redundant tree segmentation results.

Best Mask Matching: After obtaining the redundant tree segmentation results, we consider how to select the optimal tree segmentation results that correspond one-to-one with the real trees. Since SAM’s segmentation results do not contain semantic information, we consider using the automatically or manually obtained tree target detection results as the real position of the tree. We match the redundant Mask with the target detection results to obtain the tree segmentation results corresponding to the tree position. Specifically, we model this matching problem as a bipartite graph matching problem in mathematics. Bipartite graph matching, also known as bipartite matching, is a model in graph theory where given two disjoint sets of vertices A and B, and a set of edges EE \subseteq AA ×\times BB, a matching MM is a subset of EE such that no two edges share an endpoint. A maximum matching is a matching MM of maximum cardinality (maximum number of edges). This problem can be defined as:

min:iAjBcijxij\text{min:}\sum_{i\in A}\sum_{j\in B}c_{ij}x_{ij} (1)
s.t. iAxij=1, for all jB\displaystyle\sum_{i\in A}x_{ij}=1,\text{ for all }j\in B (2)
jBxij=1, for all iA\displaystyle\sum_{j\in B}x_{ij}=1,\text{ for all }i\in A
xij{0,1}, for all (i,j)E\displaystyle x_{ij}\in\{0,1\},\text{ for all }(i,j)\in E

Where the matchings are represented by variables xijx_{ij} for each edge E(i,j)E(i,j). cijc_{ij} is the cost of matching vertex i to vertex j in the bipartite graph. The constraints ensure that each vertex is matched exactly once.

In short, the bipartite graph matching problem is used to solve a set of elements in set BB that correspond one-to-one with elements in set AA. This correspondence is called maximum matching. When solving bipartite graph matching, it is necessary to define the cost of adding each edge, that is, the Cost matrix. In this article, we aim to maximize the overlap between the segmentation result and our detected target frame. We can use GIoU [Rezatofighi et al., 2019] to characterize this degree of overlap. The GIoU is defined as IoU minus the area of the smallest enclosing box that covers both bounding boxes P and G, excluding the union area of P and G and can be expressed as follows:

IoU=Area of IntersectionArea of UnionIoU=\frac{{\text{{Area of Intersection}}}}{{\text{{Area of Union}}}} (3)
GIoU=IoUAcAuAcGIoU=IoU-\frac{A_{c}-A_{u}}{A_{c}} (4)

After obtaining the Cost matrix, we use the classic Hungarian algorithm to solve bipartite graph matching and obtain the optimal segmentation result with the highest overlap with tree position.

Tree Structure Factor Estimation: After obtaining the tree segmentation results, we use the 2D segmentation results to assign labels to the point cloud within the corresponding range. This allows us to obtain the 3D single-tree segmentation results. After obtaining the 3D segmentation, we calculate the tree height, crown width, support height and crown area. We store them in the tree database together with the 2D segmentation results, external target frame and tree position information. Among them, the tree position is stored in the database in pixel coordinates. The 2D tree contour is stored in the database in MS COCO’s [Lin et al., 2014] compressed storage format.

3.2 Domain Knowledge Base

The Domain Knowledge Base is a vector database that stores embedded text of ecological knowledge in the field of forestry. The module first uses the OpenAIEmbedding API to embed the text of ecological knowledge in the forestry field and convert it into numerical vectors. The text is then segmented using the LangChain framework’s text segmentation tool with a block size of 4,000 tokens, and the resulting knowledge vectors are split into several blocks and stored in the Chroma vector database. When Tree-GPT receives a natural language instruction, it is first embedded using the OpenAIEmbedding API and converted into a numerical vector. The module then uses the Facebook AI Similarity Search (FAISS) algorithm to retrieve relevant information from the Chroma database, obtaining retrieval results denoted as SS. These results are ranked, and text blocks with a similarity greater than 0.6 are selected. These selected text blocks are then used as new inputs to the LLM using the prompt template ’Given context, could you please explain the meaning of query?’ to obtain the final inference result. This allows the model to accurately answer knowledge-based questions in the professional domain.

3.3 LLM Execution Agent

The LLM Execution Agent is a backend framework in TreeGPT that bridges the LLM’s generated code structure with the local runtime environment. Built on the LangChain toolchain, the LLM Execution Agent executes instructions in two steps: Task Planning and Code Execution. In Task Planning, the LLM decomposes natural language instructions into subtasks using prompt templates and generates corresponding content for each subtask. By breaking down tasks, the LLM is more likely to generate accurate code, as the subtasks are simpler. TreeGPT employs task planning templates, as shown in the Table 1.

Prompt
Templete
Input Output
1. Thought
Natural Language
Instruction
Task
Decomposition
2. Action
Task
Decomposition
Type of
Sub-Tasks
3. Action Input Sub-Tasks Code
4. Observation Code
Running
Results
5. Thought
Running Results +
Instruction
Output Text
6. Final Result -
Output Text+
Running Results
(Visualization, etc.)
Table 1: Process of LLM Execution Agent.

4 Experiment Results

Refer to caption
Figure 3: The results of Tree-GPT’s operation.

We use data collected from Shenzhen, Guangdong Province, China to verify the effectiveness of our method. The data was collected in 2018, with an RGB orthoimage spatial resolution of 0.025m and a Lidar resolution of 100 points per m2m^{2}. In the experiment, we cropped three 3000×\times3000 pixel areas to verify the effectiveness of the method. In terms of language model selection, we use the ChatGPT API to access the database to maximize the accuracy of code generation.

Our experiments on Tree-GPT revolve around two tasks: tree segmentation and data visualization and analysis based on natural language. The results are presented in Figure 2 and 3. As shown in Figure 2, in the tree segmentation task, Tree-GPT’s segmentation results are basically consistent with manual annotations. It can even be said that in some cases the manual annotation results may not be as accurate as SAM’s results. In addition, based on natural language-guided tree parameter query, visualization and analysis results are presented in Figures 3 respectively. The operating results of Tree-GPT are demonstrated in various tasks, including simple visualization (directly generating scatter plots), information retrieval (finding the tallest tree), complex visualization (drawing a tree growth diagram based on the position and crown width of the tree and generating a box plot of trees grouped by height), code generation, and statistical learning-based analysis and testing of tree ecological parameters (estimating the Gaussian distribution parameters of tree height and using RMSE as a confidence measure). As shown in the figure3, during the query, Tree-GPT can give correct query results and visualize them in the figure. In the data visualization task, Tree-GPT can give basically correct results in the first round. However, it may require multiple manual guidance to achieve the best visualization effect. In the data analysis task, Tree-GPT can implement simple machine learning code for various tasks.

5 Conclusion

In conclusion, Tree-GPT offers a user-friendly approach that enables real-time data updates and analysis, thereby reducing the time required for the processing workflow and allowing researchers to allocate more resources towards data interpretation and the understanding of ecological patterns. With its understanding and generation capabilities provided by Large Language Models, Tree-GPT offers a pertinent contribution to the field of tree ecological parameter extraction.

ACKNOWLEDGEMENTS

This work was supported in part by the National Key Research and Development Program of China (Project No. 2022 YFB 3903700), a Shenzhen Science and Technology Program (Project No. JCYJ 20210324093012033), the Natural Science Foundation of Guangdong Province (Project No. 2121A1515012574), the National Natural Science Foundation of China (Project Nos. 71901147, 41901329, 41971354, and 41971341,42001331), Shenzhen Key Laboratory of Digital Twin Technologies for Cities (Project No: ZDSYS20210623101800001).

References

  • [Beloiu et al., 2023] Beloiu, M., Heinzmann, L., Rehush, N., Gessler, A., Griess, V. C., 2023. Individual Tree-Crown Detection and Species Identification in Heterogeneous Forests Using Aerial RGB Imagery and Deep Learning. Remote Sensing, 15(5), 1463.
  • [Chappuis et al., 2022] Chappuis, C., Zermatten, V., Lobry, S., Le Saux, B., Tuia, D., 2022. Prompt-rsvqa: Prompting visual context to a language model for remote sensing visual question answering. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 1372–1381.
  • [Chiang et al., 2023] Chiang, W.-L., Li, Z., Lin, Z., Sheng, Y., Wu, Z., Zhang, H., Zheng, L., Zhuang, S., Zhuang, Y., Gonzalez, J. E. et al., 2023. Vicuna: An open-source chatbot impressing gpt-4 with 90%* chatgpt quality. See https://vicuna. lmsys. org (accessed 14 April 2023).
  • [Dosovitskiy et al., 2020] Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S. et al., 2020. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929.
  • [Goodfellow et al., 2016] Goodfellow, I., Bengio, Y., Courville, A., 2016. Deep learning. MIT press.
  • [He et al., 2022] He, K., Chen, X., Xie, S., Li, Y., Dollár, P., Girshick, R., 2022. Masked autoencoders are scalable vision learners. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 16000–16009.
  • [He et al., 2017] He, K., Gkioxari, G., Dollár, P., Girshick, R., 2017. Mask r-cnn. Proceedings of the IEEE international conference on computer vision, 2961–2969.
  • [He et al., 2016] He, K., Zhang, X., Ren, S., Sun, J., 2016. Deep residual learning for image recognition. Proceedings of the IEEE conference on computer vision and pattern recognition, 770–778.
  • [Indirabai et al., 2019] Indirabai, I., Nair, M. H., Nair, J. R., Nidamanuri, R. R., 2019. Estimation of forest structural attributes using ICESat/GLAS-spaceborne laser altimetry data in the western ghats region of India. Journal of Geovisualization and Spatial Analysis, 3, 1–13.
  • [Ji et al., 2023] Ji, Z., Lee, N., Frieske, R., Yu, T., Su, D., Xu, Y., Ishii, E., Bang, Y. J., Madotto, A., Fung, P., 2023. Survey of hallucination in natural language generation. ACM Computing Surveys, 55(12), 1–38.
  • [Kirillov et al., 2023] Kirillov, A., Mintun, E., Ravi, N., Mao, H., Rolland, C., Gustafson, L., Xiao, T., Whitehead, S., Berg, A. C., Lo, W.-Y. et al., 2023. Segment anything. arXiv preprint arXiv:2304.02643.
  • [LeCun et al., 2015] LeCun, Y., Bengio, Y., Hinton, G., 2015. Deep learning. nature, 521(7553), 436–444.
  • [Li et al., 2022] Li, Y., Chai, G., Wang, Y., Lei, L., Zhang, X., 2022. Ace r-cnn: An attention complementary and edge detection-based instance segmentation algorithm for individual tree species identification using uav rgb images and lidar data. Remote Sensing, 14(13), 3035.
  • [Lin et al., 2017] Lin, T.-Y., Goyal, P., Girshick, R., He, K., Dollár, P., 2017. Focal loss for dense object detection. Proceedings of the IEEE international conference on computer vision, 2980–2988.
  • [Lin et al., 2014] Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Dollár, P., Zitnick, C. L., 2014. Microsoft coco: Common objects in context. Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13, Springer, 740–755.
  • [Lobry et al., 2020] Lobry, S., Marcos, D., Murray, J., Tuia, D., 2020. RSVQA: Visual question answering for remote sensing data. IEEE Transactions on Geoscience and Remote Sensing, 58(12), 8555–8566.
  • [Lutz et al., 2012] Lutz, J. A., Larson, A. J., Swanson, M. E., Freund, J. A., 2012. Ecological Importance of Large-Diameter Trees in a Temperate Mixed-Conifer Forest. PLOS ONE, 7(5), 1-15. https://doi.org/10.1371/journal.pone.0036131.
  • [OpenAI, 2023] OpenAI, 2023. Gpt-4 technical report.
  • [Ouyang et al., 2022] Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C., Mishkin, P., Zhang, C., Agarwal, S., Slama, K., Ray, A. et al., 2022. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35, 27730–27744.
  • [Qian et al., 2023] Qian, C., Yao, C., Ma, H., Xu, J., Wang, J., 2023. Tree species classification using airborne lidar data based on individual tree segmentation and shape fitting. Remote Sensing, 15(2), 406.
  • [Rezatofighi et al., 2019] Rezatofighi, H., Tsoi, N., Gwak, J., Sadeghian, A., Reid, I., Savarese, S., 2019. Generalized intersection over union: A metric and a loss for bounding box regression. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 658–666.
  • [Strîmbu and Strîmbu, 2015] Strîmbu, V. F., Strîmbu, B. M., 2015. A graph-based segmentation algorithm for tree crown extraction using airborne LiDAR data. ISPRS Journal of Photogrammetry and Remote Sensing, 104, 30–43.
  • [Sun et al., 2019] Sun, Y., Xin, Q., Huang, J., Huang, B., Zhang, H., 2019. Characterizing tree species of a tropical wetland in southern china at the individual tree level based on convolutional neural network. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 12(11), 4415–4425.
  • [Suratno et al., 2009] Suratno, A., Seielstad, C., Queen, L., 2009. Tree species identification in mixed coniferous forest using airborne laser scanning. ISPRS Journal of Photogrammetry and Remote Sensing, 64(6), 683–693.
  • [Tochon et al., 2015] Tochon, G., Feret, J.-B., Valero, S., Martin, R. E., Knapp, D. E., Salembier, P., Chanussot, J., Asner, G. P., 2015. On the use of binary partition trees for the tree crown segmentation of tropical rainforest hyperspectral images. Remote sensing of environment, 159, 318–331.
  • [Turner et al., 2006] Turner, J. A., Buongiorno, J., Zhu, S., 2006. An economic model of international wood supply, forest stock and forest area change. Scandinavian Journal of Forest Research, 21(1), 73–86.
  • [Vaswani et al., 2017] Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, Ł., Polosukhin, I., 2017. Attention is all you need. Advances in neural information processing systems, 30.
  • [Wagner et al., 2018] Wagner, F. H., Ferreira, M. P., Sanchez, A., Hirye, M. C., Zortea, M., Gloor, E., Phillips, O. L., de Souza Filho, C. R., Shimabukuro, Y. E., Aragão, L. E., 2018. Individual tree crown delineation in a highly diverse tropical forest using very high resolution satellite images. ISPRS journal of photogrammetry and remote sensing, 145, 362–377.
  • [Weinstein et al., 2019] Weinstein, B. G., Marconi, S., Bohlman, S., Zare, A., White, E., 2019. Individual tree-crown detection in RGB imagery using semi-supervised deep learning neural networks. Remote Sensing, 11(11), 1309.
  • [Yang et al., 2020] Yang, J., Kang, Z., Cheng, S., Yang, Z., Akwensi, P. H., 2020. An individual tree segmentation method based on watershed algorithm and three-dimensional spatial distribution analysis from airborne LiDAR point clouds. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 13, 1055–1067.
  • [Yang et al., 2019] Yang, Q., Su, Y., Jin, S., Kelly, M., Hu, T., Ma, Q., Li, Y., Song, S., Zhang, J., Xu, G. et al., 2019. The influence of vegetation characteristics on individual tree segmentation methods with airborne LiDAR data. Remote Sensing, 11(23), 2880.
  • [Zhang et al., 2023] Zhang, C., Liu, L., Cui, Y., Huang, G., Lin, W., Yang, Y., Hu, Y., 2023. A Comprehensive Survey on Segment Anything Model for Vision and Beyond. arXiv preprint arXiv:2305.08196.
  • [Zhang et al., 2022] Zhang, C., Zhou, J., Wang, H., Tan, T., Cui, M., Huang, Z., Wang, P., Zhang, L., 2022. Multi-species individual tree segmentation and identification based on improved mask R-CNN and UAV imagery in mixed forests. Remote Sensing, 14(4), 874.
  • [Zhao et al., 2023] Zhao, W. X., Zhou, K., Li, J., Tang, T., Wang, X., Hou, Y., Min, Y., Zhang, B., Zhang, J., Dong, Z. et al., 2023. A survey of large language models. arXiv preprint arXiv:2303.18223.
  • [Zhu et al., 2023] Zhu, D., Chen, J., Shen, X., Li, X., Elhoseiny, M., 2023. Minigpt-4: Enhancing vision-language understanding with advanced large language models. arXiv preprint arXiv:2304.10592.