This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Towards Lean Research Inception: Assessing Practical Relevance of Formulated Research Problems

Anrafel Fernandes Pereira PUC-RioRio de JaneiroBrazil afpereira@inf.puc-rio.br Marcos Kalinowski PUC-RioRio de JaneiroBrazil kalinowski@inf.puc-rio.br Maria Teresa Baldassarre University of BariBariItaly mariateresa.baldassarre@uniba.it Jürgen Börstler Blekinge Institute of TechnologyKarlskronaSweden jurgen.borstler@bth.se Nauman bin Ali Blekinge Institute of TechnologyKarlskronaSweden nauman.ali@bth.se  and  Daniel Mendez Blekinge Institute of TechnologyKarlskronaSweden daniel.mendez@bth.se
(2025)
Abstract.

[Context] The lack of practical relevance in many Software Engineering (SE) research contributions is often rooted in oversimplified views of industrial practice, weak industry connections, and poorly defined research problems. Clear criteria for evaluating SE research problems can help align their value, feasibility, and applicability with industrial needs. [Goal] In this paper, we introduce the Lean Research Inception (LRI) framework, designed to support the formulation and assessment of practically relevant research problems in SE. We describe its initial evaluation strategy conducted in a workshop with a network of SE researchers experienced in industry-academia collaboration and report the evaluation of its three assessment criteria (valuable, feasible, and applicable) regarding their importance and completeness in assessing practical relevance. [Method] We applied LRI retroactively to a published research paper, engaging workshop participants in discussing and assessing the research problem by applying the proposed criteria using a semantic differential scale. Participants provided feedback on the criteria’s importance and completeness, drawn from their own experiences in industry-academia collaboration. [Results] The findings reveal an overall agreement on the importance of the three criteria – valuable (83.3%), feasible (76.2%), and applicable (73.8%) – for aligning research problems with industrial needs. Qualitative feedback suggested adjustments in terminology with a clearer distinction between feasible and applicable, and refinements for valuable by more clearly considering business value, ROI, and originality. [Conclusion] While LRI still constitutes ongoing research and requires further evaluation, our emerging results strengthen our confidence that the three criteria applied using the semantic differential scale can already help the community assess the practical relevance of SE research problems.

Relevance Assessment, Research Problem Formulation, Practical Relevance
copyright: acmlicensedjournalyear: 2025doi: XXXXXXX.XXXXXXXconference: The 29th International Conference on Evaluation and Assessment in Software Engineering; 17–20 June, 2025; Istanbul, Türkiyeccs: Software and its engineering Development frameworks and environmentsccs: General and reference Empirical studiesccs: Social and professional topics Computing industry

1. Introduction

The practical relevance of Software Engineering (SE) research is often limited by a disconnect between academia and industry, with studies failing to address practically relevant challenges despite their theoretical contributions (Garousi et al., 2020; Franch et al., 2020; Winters, 2024). Gorschek et al. (Gorschek et al., 2006; Gorschek and Mendez, 2021) emphasize that a well-defined problem formulation is critical for research to generate industrial impact. Many research problems are formulated in isolation, lacking input from industrial practice (Gorschek and Mendez, 2021). This disconnect hinders understanding of real and practical industry challenges. Ensuring practical relevance requires structured approaches that align research with industry needs through clear problem formulation and assessment (Ali, 2016; Molléri et al., 2023; Storey et al., 2024).

To integrate academic and industry perspectives early in the research process, we introduce the Lean Research Inception (LRI) framework to support the formulation and initial assessment of practically relevant SE research problems. In this paper, we present the complete vision of the LRI framework and an evaluation of its research problem assessment phase. The assessment employs a semantic differential scale (Heise, 1970) to assess the relevance of formulated research problems based on three criteria: valuable (whether solving the problem can generate meaningful impact for industrial practice), applicable (whether the problem leads to practical and usable outcomes in industry), and feasible (whether the problem can realistically be investigated given available resources). The scale was inspired by agile principles and Lean Startup’s Minimum Viable Product (MVP) concept (Ries, 2011), aiming to enhance the effectiveness of aligning research with industry needs.

The empirical evaluation was conducted through a workshop with 42 senior SE researchers at the 2024 annual meeting of the International Software Engineering Research Network (ISERN111https://isern.iese.de/). The study focused on evaluating the three key criteria – valuable, feasible, and applicable – used in the semantic differential scale. Participants analyzed these criteria by discussing an example research problem, capturing their insights through collaborative annotations, and individually applying the criteria to assess the problem’s relevance. Following this discussion, they completed a survey on the perceived importance and completeness of each criterion. The survey provided quantitative and qualitative insights into how well these criteria are perceived as important to support an initial assessment of the practical relevance of formulated research problems.

The results indicate that our criteria, valuable (83.3%), feasible (76.2%), and applicable (73.8%) are appropriate for assessing research problem relevance, with high levels of agreement among participants. Furthermore, qualitative feedback suggested adjustments in terminology for a clearer distinction between feasible and applicable, and refinements for valuable to include concrete examples to ease understanding, such as business value, ROI, and originality.

2. Related Work

Software Engineering research is primarily aimed at industrial applications, which makes alignment with industry needs crucial for practical relevance (Winters, 2024). Achieving this requires strategies that balance theoretical advancements with real-world applicability (Stol and Fitzgerald, 2018). Ivarsson et al. (Ivarsson and Gorschek, 2011) proposed a model to assess the rigor and industrial relevance of technology evaluations in SE. In their analyses, the model revealed that most evaluations lacked both rigor and industrial relevance. Additionally, the study found no significant improvement in industrial relevance over time.

Garousi et al. (Garousi et al., 2020) investigate the practical relevance of SE research, synthesizing community opinions and evidence through a multivocal literature review (MLR). They identify three key factors limiting relevance: too simplistic views of practice, weak industry connections, and poor problem identification. To address this, they recommend adopting appropriate research approaches, focusing on practical problems, and strengthening industry collaboration. The study emphasizes the need for rigorous empirical research to better align SE studies with industrial needs. Winters (Winters, 2024) further highlights the misalignment between SE research and industry needs, criticizing the lack of practical context, narrow problem scopes, and limited scalability, emphasizing the need for research that delivers measurable and practical value to industry.

Petersen et al. (Petersen et al., 2024) propose a reasoning framework to enhance the design, assessment, and reporting of industrial relevance in SE research. Given the lack of consensus on defining and measuring relevance, they review key attributes such as applicability, context, and practical impact. Their framework, structured around six aspects (what, how, where, who, when, and why), provides an approach for evaluating and communicating research relevance of applied research in industry contexts. Storey et al. (Storey et al., 2024) propose a playbook for researching disruptive innovations in SE. They emphasize interdisciplinary collaboration, industry engagement, and iterative experimentation while underscoring the need for clear criteria to assess impact and applicability in industrial and social contexts.

These studies emphasize the importance of aligning SE research with industry needs, reinforcing the motivations behind our work. Ivarsson et al. (Ivarsson and Gorschek, 2011) highlight the persistent lack of rigor and industrial relevance in SE evaluations, underscoring the need for structured approaches to bridge this gap. Garousi et al. (Garousi et al., 2020) and Winters (Winters, 2024) critique the disconnect between academia and industry, pointing to weak collaboration and impractical research focus—challenges LRI directly addresses by integrating practitioners early in the research process. Petersen et al. (Petersen et al., 2024) propose a reasoning framework for industrial relevance, which helped to define LRI’s problem formulation attributes and assessment criteria. Finally, Storey et al.(Storey et al., 2024) provide valuable insights into assessing the impact of disruptive innovations.

LRI advances previous work by proposing an iterative framework that integrates academic and industry perspectives from the outset to support the formulation of research problems and an initial assessment of their practical relevance through clear and easy-to-apply criteria.

3. Lean Research Inception

The Lean Research Inception (LRI) framework was developed to help bridging the gap between SE research and industrial practice. Grounded in agile principles and methodologies such as Design Thinking (Plattner et al., 2009), Lean Startup (Ries, 2011), and Lean Inception (Caroli, 2017), LRI fosters collaboration between researchers and practitioners from the very beginning of a research project. Its core objective is to integrate these stakeholders early on to ensure that research problems are practically relevant and applicable to real-world scenarios.

LRI complements the technology transfer model proposed by Gorschek et al. (Gorschek et al., 2006), which defines seven stages for transitioning research into industrial practice. LRI focuses on how to engage in one of the most critical stages: “Problem Formulation”. LRI refines this stage with a structured and transparent process. It includes practical recommendations (Garousi et al., 2016) to promote early alignment between academia and industry. The approach actively involves SE researchers and practitioners from the beginning. However, it acknowledges that engaging practitioners is challenging. By promoting this collaborative approach, LRI aims to enhance the relevance and applicability of SE research, fostering a more effective exchange between academia and industry.

The LRI framework comprises five sequential phases, as illustrated in Figure 1. We created an interactive board template that can be used to guide these phases and be filled out collaboratively in a shared workspace. Our open science repository (Fernandes Pereira et al., 2025) provides a blank template that allows visualizing the framework in detail and a complete example filled out based on a published research paper.

Phase 1 - Problem Vision Outline: SE researchers work collaboratively to create an initial draft of the practical problem by filling attributes of the “Problem Vision” board (problem outline (what/how/why), context (where/when), implications/impacts (why), practitioners (who), evidence (how), objective (what/how), and research questions (what)). The attributes were selected based on previous research on relevance in SE, also considering their support in the broader literature on SE, especially regarding their role in problem formulation (Garousi et al., 2020; Gorschek et al., 2006; Ivarsson and Gorschek, 2011; Petersen et al., 2024).

Phase 2 - Problem Vision Alignment: Practitioners join the SE researchers in a collaborative workshop. After introductions, the researchers present the initial “Problem Vision”. The participants then review, discuss, and refine its seven attributes to ensure a shared understanding and improve the problem formulation.

Phase 3 - Research Problem Formulation: Researchers and practitioners refine and document the research problem based on the discussions in Phase 2. The formulated research problem serves as the basis for the assessment in the next phase.

Phase 4 - Research Problem Assessment: Participants individually assess the relevance of the formulated research problem using a semantic differential scale (Heise, 1970). Adapted to LRI, this scale assesses three key criteria of practical relevance in SE, aligned with agile principles and the concept of a Minimum Viable Product (MVP) (Ries, 2011). They are: “Worthless - Valuable”, which assesses whether solving the problem can generate meaningful impact for industrial practice; “Infeasible - Feasible”, which assesses whether the problem can realistically be investigated given available resources; and “Inapplicable - Applicable”, which assesses if the problem leads to practical and usable outcomes for the industry. Each criterion is rated on a seven-point scale, enabling a detailed analysis.

Refer to caption
Figure 1. Lean Research Inception Overview

Phase 5 - Go/Pivot/Abort Decision: In this final phase, the group assessment results are consolidated. Aligned with the MVP-approach (Ries, 2011), LRI seeks to ensure that the research problem is relevant (valuable, feasible, and applicable) to real-world SE scenarios. Similarly to MVPs, if the perceived relevance is high, the investigation should continue (e.g., to the ‘Candidate Solution’ stage of the Gorschek et al.’s model (Gorschek et al., 2006)). If the perceived relevance is low but adjustable, the research problem should be realigned (or pivoted), restarting at Phase 2. However, if the perceived relevance is low and cannot be fixed, the investigation should be aborted to prevent wasted effort on irrelevant research.

4. Evaluation

We conducted an empirical study with 42 senior SE researchers experienced in industry-academia collaboration to evaluate three criteria—valuable, feasible, and applicable—in assessing the practical relevance of research problems. As part of our initial evaluation strategy, we enacted the LRI framework within the ISERN workshop. Participants were introduced to a pre-filled “Problem Vision” board, based on a published research problem (Cabral et al., 2024), to explore its structure and attributes (Phase 1). They then discussed and refined the problem collaboratively (Phases 2 and 3), followed by an individual assessment using the semantic differential scale (Phase 4). The exercise concluded with the consolidation of perceptions and submission of individual feedback via a survey (Phase 5).

This study offers a complete view of the LRI framework, with a specific focus on Phase 4, which evaluates the practical relevance of formulated research problems. The analysis is based on data collected from survey responses on the perceived importance and completeness of the three evaluation criteria. While this represents a small-scale evaluation rather than a full case study (Wohlin and Rainer, 2022), this approach tends to be well-suited for the initial assessment of the prospective value of a new proposal (Wohlin and Rainer, 2022). We adhered to the case study reporting structure recommended by Runeson et al.(Runeson et al., 2012).

4.1. Goal and Research Questions

We define the goal of this study using the goal definition template of the Goal Question Metric (GQM) paradigm (Basili and Rombach, 1988) as follows:

Analyze the criteria (valuable, feasible, and applicable) of the semantic differential scale for the purpose of characterization with respect to the perception of their importance and completeness to support practical relevance assessment of formulated research problems from the point of view of senior SE researchers with experience in industry-academia collaborations in the context of conducting a practical relevance assessment and providing feedback on the importance of the criteria.”

Therefrom, we formulate the following two research questions:

  • RQ1: To what degree are the three criteria perceived as important to support practical relevance assessment of formulated research problems?

  • RQ2: What other criteria are perceived as important to support practical relevance assessment of formulated research problems?

4.2. Case and Subject Selection

We conducted this study during an in-person workshop at ISERN 2024 in Barcelona, Spain, engaging 42 senior SE researchers with experience in industry-academia collaboration. The setting provided a valuable opportunity to explore our objectives by leveraging participants’ expertise in a collaborative environment. We retrospectively applied the LRI phases using a research problem from a published article (Cabral et al., 2024). We selected the research problem from this paper as the discussion case because it explores the application of familiar software design principles to AI-enabled systems and was recognized with an award for its practical relevance (Staron et al., 2024).

4.3. Instrumentation

We carefully designed and reviewed all materials for this study to ensure the reliability of data collection. To support this process, we conducted a pilot workshop with twelve master’s and Ph.D. students experienced in Experimental and Empirical SE at ExACTa222https://www.exacta.inf.puc-rio.br PUC-Rio. Their familiarity with empirical research enabled them to provide valuable feedback on the clarity, consistency, and usability of the artifacts before their use in the main study.

The study instrumentation consisted of a slide presentation and three main artifacts structured in three envelopes each corresponding to a specific activity, to guide participants through a step-by-step collaborative process. The first envelope included a pre-filled A2-format “Problem Vision” board and the article by Cabral et al. (Cabral et al., 2024), which participants reviewed and discussed in the first activity, documenting their insights on the board using post-its. The second envelope included a semantic differential scale (Heise, 1970) to assess the previously discussed research problem based on value, applicability, and feasibility, which participants applied during the second activity. The third envelope included a Likert-type scale survey with optional open-ended questions to evaluate the importance and completeness of LRI’s attributes and criteria, and to gather suggestions for additional elements not covered by the framework. Participants were divided into five heterogeneous groups to ensure diverse perspectives and balanced discussions. Each group received the three envelopes. All materials are anonymously available in our open science repository (Fernandes Pereira et al., 2025).

4.4. Data Collection and Analysis Procedures

We collected both quantitative and qualitative feedback using the outlined artifacts. Our mixed-methods approach combined statistical analysis of Likert-type scale responses with qualitative analysis of open-ended feedback to comprehensively assess participant perceptions. For the quantitative analysis, we calculated agreement level frequencies for each criterion. For the qualitative analysis, we examined open-ended responses to support the quantitative findings, capture nuances, and identify divergences. All data was recorded in a spreadsheet, available in our open science repository (Fernandes Pereira et al., 2025).

5. Results

5.1. Case and Subject Description

The evaluation was carried out through a 90-minute collaborative workshop at ISERN 2024, structured into five steps: (i) an introductory presentation (20 min); (ii) group formation and material distribution (5 min); (iii) discussion of the research problem using the “Problem Vision” board (35 min); (iv) assessment of the relevance of the formulated problem using the semantic differential scale (15 min); and (v) a survey (15 min).

5.2. Perceived importance of criteria (RQ1)

The quantitative analysis of the collected data, shown in Figure 2, revealed a general trend indicating the importance of the three criteria as dimensions of the semantic differential scale. In the following, we present a detailed criterion-by-criterion analysis.

Refer to caption
Figure 2. Questionnaire results

Valuable: The results for the valuable criterion demonstrated acceptance, with 83.3% of participants agreeing to some extent with its importance (21.4% Slightly Agree and 61.9% Agree). Only 4.8% expressed some level of disagreement (Disagree and Slightly Disagree), while 11.9% remained neutral (Neither Agree nor Disagree). Qualitative feedback reinforces this perception. Participants emphasized that research should focus on meaningful and practical problems. P15 noted, “Research can be done that no one cares about; the research should be conducted with a salient problem of stakeholders,” highlighting the importance of addressing real-world issues. Similarly, P35 stated, “No point in doing something with no utility,” reinforcing the need for research to provide tangible benefits.

Despite this consensus, some participants suggested refinements. P11 recommended reconsidering the term “Worthless” in the scale, stating, “Lower end of the scale ‘Worthless’ sounds too negative. Consider an alternative word?” P16 suggested a clearer definition: “I suggest providing a definition of the criteria,” while P27 proposed breaking it down into specific elements: “I would suggest breaking down the criterion into elements that can be used to make a useful decision.” Further concerns about subjectivity emerged among neutral and disagreeing participants. P28 described the criterion as “very subjective,” while P20 found it “too fuzzy,”, indicating a need for more precise wording and refinements to improve clarity.

Feasible: The feasible criterion also received mainly positive results, with 76.2% of participants agreeing to some extent with its importance (16.7% Slightly Agree and 59.5% Agree). Meanwhile, 16.6% remained neutral, and 7.2% expressed some level of disagreement (2.4% Slightly Disagree and 4.8% Disagree). Qualitative feedback highlighted the importance of feasibility for applied research. P25 stated, “Key for applied research,” and P15 emphasized feasibility’s influence on early investigative stages: “The feasibility can influence the ‘investigation’ step incrementally.”Nonetheless, some participants reported difficulties in fully assessing the criterion. For example, P37 noted, “It is important but I missed information to answer it,”. P40 questioned, “From research or industry perspective?” suggesting that the research perspective should be clearer stated. Those who were neutral or slightly disagreed expressed difficulties in applying the criterion. P17 mentioned the need for clarifying available resources: “Clarify available resources.” P39 raised difficulties in assessing feasibility during problem definition: “Is it possible to assess this dimension at the moment of problem definition?”

Applicable: The results for applicable almost mirrored those of feasible. Approximately 73.8% of participants agreed with its importance (14.3% Slightly Agree and 59.5% Agree), while 16.6% were neutral, and 9.6% expressed some level of disagreement (4.8% Slightly Disagree and 4.8% Disagree). Participants who agreed emphasized its importance while noting limitations in its assessment. P24 commented, “Important, but the scale is not helpful,” while P06 observed that feedback provided for other criteria also applied to Applicable. Neutral participants raised concerns about its clarity and practicality. For example, P35 stated, “This maps to it being useful,” while P39 questioned, “I wonder how easy it is to assess this dimension at the moment of defining the problem.” Disagreeing participants raised foundational issues with the criterion. P20 described it as “Too fuzzy,” and P13 noted that applicability could only be determined after stakeholders are convinced of a problem’s relevance: “Applicability is independent of relevance. Deciding applicability only be possible once a company was convinced of the problem’s relevance.” Additionally, P31 observed a strong correlation between feasible and applicable: “They are very correlated. If something is unfeasible, it would be directly inapplicable.”. This indicates an opportunity to provide a clearer distinction between these two criteria.

Overall, the quantitative and qualitative findings confirm that all three criteria—valuable, feasible, and applicable—are mainly perceived as important for the assessment of research problem relevance. However, they also highlight opportunities for refinement, particularly in improving the clarity and contextualization of the criteria to enhance their effectiveness in future evaluations.

5.3. Completeness of criteria (RQ2)

To answer RQ2, participants suggested additional criteria they consider important for assessing practical relevance. The following analysis summarizes their input.

Business value and technical impact: Participants P19 and P20 emphasized the importance of distinguishing between business value and technical impact. While aligned with the valuable criterion, their suggestions highlight the need to evaluate economic and strategic benefits separately from technical contributions.

Measurable benefits: P14 proposed including “expected measurable benefit,” which resonates with the valuable criterion. This suggestion adds depth by emphasizing the need for clear, quantifiable outcomes, potentially improving the precision with which value is assessed.

Originality: P40 suggested assessing originality to ensure research problems explore new areas. Although not explicitly covered in the current criteria, originality complements valuable by emphasizing innovation, helping to distinguish impactful problems from those that replicate existing solutions.

Cost and Return on Investment (ROI): P24 and P35 highlighted cost and ROI as essential factors. This economic perspective partially aligns with feasible and partially with valuable, emphasizing the need to consider resource efficiency explicitly.

Impact and limitations: P23 and P37 emphasized the importance of assessing both the impact and limitations of solutions. While aligning with valuable and feasible, these factors help balance potential benefits with implementation challenges. P37 specifically highlighted the need to clarify limitations to better assess feasibility, stating, “I think that main limitations/restrictions should be made clear in order to reflect on feasibility.”

Applicability requirements: P28 proposed defining “Requirements for the applicability”, suggesting defining specific thresholds to make applicable more measurable.

Problem importance: P07 questioned: “Is this a problem worth putting resources into?”. This consideration aligns with valuable and feasible, balancing problem importance with the resources needed for its resolution.

Intention to use: P25 suggested assessing intention to use, based on the Technology Acceptance Model. This recommendation further details applicable by emphasizing end-user adoption.

Overall, the additional recommendations overlap or refine existing criteria.Valuable may consider business value, measurable benefits, originality, ROI, and problem importance; feasible involves analyzing costs and limitations; and applicable gains specificity through applicability requirements and intention to use.

6. Discussion

Quantitative results showed high levels of agreement (slightly agree and agree) on the importance of the criteria, especially for valuable (83.3%), followed by feasible (76.2%) and applicable (73.8%). Hence, our initial results highlight the importance of the semantic differential scale criteria (valuable, feasible, applicable) for an initial assessment of research relevance. Qualitative feedback refined these findings, suggesting improvements in terminology definitions. For instance, the participants suggested a clearer distinction between feasible and applicable. This could be solved by better explaining that feasibility takes the research perspective, while applicability considers the industry perspective.

Participants suggested refinements rather than entirely new criteria, focusing on aspects of valuable (e.g., business value, measurable benefits, originality, ROI, problem importance), feasible (e.g., costs, limitations), and applicable (e.g., applicability requirements, intention to use). These suggestions reflect a need to clarify value, add contextual detail for feasibility, and define measurable applicability criteria. However, as noted by some participants, evaluating these dimensions precisely during problem formulation is difficult. Thus, we consider the MVP-inspired assessment scale suitable for early-stage evaluation, enabling flexible, collective judgment before more detailed metrics are applied.

The contribution of this work complements related work focusing on assessing the practical relevance of research problems in SE. While Garousi et al. (Garousi et al., 2020) highlight gaps in practical relevance and Winters (Winters, 2024) criticizes the limited applicability of academic studies, LRI addresses these concerns by integrating practitioners early in the process of problem formulation and relevance assessment. In line with the need of assessing relevance described by Ivarsson et al. (Ivarsson and Gorschek, 2011), and differently from the more complete reasoning framework to assess industrial relevance (Petersen et al., 2024), inspired by the desired properties of MVP (Ries, 2011), LRI defines a simple semantic differential scale with three main criteria –valuable, feasible and applicable– that can be used to support an initial collective assessment of the relevance of formulated research problems.

7. Threats to Validity

We addressed the four categories of validity threats described by Wohlin et al. (Wohlin et al., 2024):

Internal Validity: To mitigate threats to internal validity, we randomly assigned participants to groups, and each participant individually applied the assessment scale for the selected case before answering the survey questions on the assessment criteria. Furthermore, the survey provided open-ended questions for justification and allowed participants to suggest additional criteria.

External Validity: Conducting the study with a group of participants that may limit generalizability beyond (senior) SE researchers. However, we argue that ISERN members, as per ISERN’s mission, are expected to be experts in empirical SE and industry-academia collaboration, and thus render this group highly relevant for this study. Hence, we consider these emerging results to be useful initial indications. Nevertheless, future research should involve industry stakeholders for broader generalizability.

Construct Validity: The Likert-type scale may not have fully captured perceptions, and interpretations of the criteria could vary. We addressed this by including open-ended questions, carefully reviewing workshop materials, and conducting a pilot study with master and Ph.D. students to refine study instruments.

Conclusion Validity: The sample size (42 participants) and reliance on Likert-type scale responses limit inferential statistics. To strengthen the findings, we triangulated quantitative data with qualitative analysis, ensuring a more comprehensive interpretation.

8. Concluding Remarks

This study introduced the Lean Research Inception (LRI) framework as a structured approach to support the formulation and initial assessment of practically relevant research problems in SE. We evaluated the three criteria of LRI’s Semantic Differential Scale in terms of importance and completeness to support this initial assessment. Our findings indicate an overall agreement on the importance of the three criteria – valuable (83.3%), feasible (76.2%), and applicable (73.8%) – in aligning research problems with industrial needs. Qualitative feedback suggested refining terminology to better distinguish between feasible and applicable and expanding valuable to encompass business value, ROI, and originality.

Although LRI remains an ongoing research effort requiring further evaluation, these emerging results indicate that early practitioner involvement and the use of the semantic differential scale offer a practical approach for the initial assessment of the relevance of SE research problems. Future research could refine the terminology of the scale and include practical examples for each dimension. In addition, more industry-driven case studies are needed to evaluate the effectiveness of LRI in real-world collaboration contexts.

Acknowledgements.
We thank the 42 SE researchers from ISERN who voluntarily participated in our study. This work has been partially supported by the following projects: SERICS, grant PE00000014, MUR National Recovery and Resilience Plan, European Union - NextGenerationEU (EU-NGEU); QUASAR, grant 2022T2E39C, PRIN 2022 MUR program, EU-NGEU; S.E.R.T. Research Profile project, KKS foundation; CNPq, grant 312275/2023-4; FAPERJ, grant E-26/204.256/2024; and CAPES, finance code 001.

References

  • (1)
  • Ali (2016) Nauman Bin Ali. 2016. Is effectiveness sufficient to choose an intervention?: Considering resource use in empirical software engineering. In Proceedings of the 10th ACM/IEEE International Symposium on Empirical Software Engineering and Measurement. 54:1–54:6. doi:10.1145/2961111.2962631
  • Basili and Rombach (1988) Victor R Basili and H Dieter Rombach. 1988. The TAME project: Towards improvement-oriented software environments. IEEE Transactions on software engineering 14, 6 (1988), 758–773.
  • Cabral et al. (2024) Raphael Cabral, Marcos Kalinowski, Maria Teresa Baldassarre, Hugo Villamizar, Tatiana Escovedo, and Hélio Lopes. 2024. Investigating the Impact of SOLID Design Principles on Machine Learning Code Understanding. In Proceedings of the IEEE/ACM 3rd International Conference on AI Engineering-Software Engineering for AI. 7–17.
  • Caroli (2017) Paulo Caroli. 2017. Lean inception. São Paulo, BR: Caroli. org (2017).
  • Fernandes Pereira et al. (2025) Anrafel Fernandes Pereira, Marcos Kalinowski, Maria Teresa Baldassarre, Jürgen Börstler, Nauman bin Ali, and Daniel Mendez. 2025. Artifacts - Towards Lean Research Inception: Assessing Practical Relevance of Formulated Research Problems. doi:10.5281/zenodo.14989559
  • Franch et al. (2020) Xavier Franch, Daniel Mendez, Andreas Vogelsang, Rogardt Heldal, Eric Knauss, Marc Oriol, Guilherme H Travassos, Jeffrey C Carver, and Thomas Zimmermann. 2020. How do practitioners perceive the relevance of requirements engineering research? IEEE Transactions on Software Engineering 48, 6 (2020), 1947–1964.
  • Garousi et al. (2020) Vahid Garousi, Markus Borg, and Markku Oivo. 2020. Practical relevance of software engineering research: synthesizing the community’s voice. Empirical Software Engineering 25 (2020), 1687–1754.
  • Garousi et al. (2016) Vahid Garousi, Kai Petersen, and Baris Ozkan. 2016. Challenges and best practices in industry-academia collaborations in software engineering: A systematic literature review. Information and Software Technology 79 (2016), 106–127.
  • Gorschek et al. (2006) Tony Gorschek, Per Garre, Stig Larsson, and Claes Wohlin. 2006. A model for technology transfer in practice. IEEE software 23, 6 (2006), 88–95.
  • Gorschek and Mendez (2021) Tony Gorschek and Daniel Mendez. 2021. Solving Problems or Enabling Problem-Solving? From Purity in Empirical Software Engineering to Effective Co-production (Invited Keynote). In Software Quality: Future Perspectives on Software Engineering Quality: 13th International Conference. Springer, Cham, Switzerland, 109–116. doi:10.1007/978-3-030-65854-0_9
  • Heise (1970) David R Heise. 1970. The semantic differential and attitude research. Attitude measurement 4 (1970), 235–253.
  • Ivarsson and Gorschek (2011) Martin Ivarsson and Tony Gorschek. 2011. A method for evaluating rigor and industrial relevance of technology evaluations. Empirical Software Engineering 16 (2011), 365–395.
  • Molléri et al. (2023) Jefferson Seide Molléri, Emilia Mendes, Kai Petersen, and Michael Felderer. 2023. Determining a core view of research quality in empirical software engineering. Computer Standards & Interfaces 84 (2023), 103688.
  • Petersen et al. (2024) Kai Petersen, Jürgen Börstler, Nauman Bin Ali, and Emelie Engström. 2024. Revisiting the construct and assessment of industrial relevance in software engineering research. In Proceedings of the 1st IEEE/ACM International Workshop on Methodological Issues with Empirical Studies in Software Engineering. 17–20.
  • Plattner et al. (2009) Hasso Plattner, Christoph Meinel, and Ulrich Weinberg. 2009. Design thinking. Springer.
  • Ries (2011) Eric Ries. 2011. The lean startup: How today’s entrepreneurs use continuous innovation to create radically successful businesses. Crown Currency.
  • Runeson et al. (2012) Per Runeson, Martin Host, Austen Rainer, and Bjorn Regnell. 2012. Case study research in software engineering: Guidelines and examples. John Wiley & Sons.
  • Staron et al. (2024) Miroslaw Staron, Silvia Abrahão, Grace Lewis, Henry Muccini, and Chetan Honnenahalli. 2024. Bringing Software Engineering Discipline to the Development of AI-Enabled Systems. IEEE Software 41, 5 (2024), 79–82.
  • Stol and Fitzgerald (2018) Klaas-Jan Stol and Brian Fitzgerald. 2018. The ABC of software engineering research. ACM Transactions on Software Engineering and Methodology 27, 3 (2018), 1–51.
  • Storey et al. (2024) Margaret-Anne Storey, Daniel Russo, Nicole Novielli, Takashi Kobayashi, and Dong Wang. 2024. A disruptive research playbook for studying disruptive innovations. ACM Transactions on Software Engineering and Methodology 33, 8 (2024), 1–29.
  • Winters (2024) Titus Winters. 2024. Thoughts on applicability. Journal of Systems and Software 215 (2024), 112086.
  • Wohlin and Rainer (2022) Claes Wohlin and Austen Rainer. 2022. Is it a case study?—A critical analysis and guidance. Journal of Systems and Software 192 (2022), 111395.
  • Wohlin et al. (2024) Claes Wohlin, Per Runeson, Martin Höst, Magnus C. Ohlsson, Björn Regnell, and Anders Wesslén. 2024. Experimentation in Software Engineering, Second Edition. Springer.