Persona:
Fernández Amoros, David José

Cargando...
Foto de perfil
Dirección de correo electrónico
ORCID
0000-0003-3758-0195
Fecha de nacimiento
Proyectos de investigación
Unidades organizativas
Puesto de trabajo
Apellidos
Fernández Amoros
Nombre de pila
David José
Nombre

Resultados de la búsqueda

Mostrando 1 - 10 de 10
  • Publicación
    A scalable approach to exact model and commonality counting for extended feature models.
    (Institute of Electrical and Electronics Engineers (IEEE), 2014-05-29) Fernández Amoros, David José; Heradio Gil, Rubén; Cerrada Somolinos, José Antonio; Cerrada Somolinos, Carlos
    A software product line is an engineering approach to efficient development of software product portfolios. Key to the success of the approach is to identify the common and variable features of the products and the interdependencies between them, which are usually modeled using feature models. Implicitly, such models also include valuable information that can be used by economic models to estimate the payoffs of a product line. Unfortunately, as product lines grow, analyzing large feature models manually becomes impracticable. This paper proposes an algorithm to compute the total number of products that a feature model represents and, for each feature, the number of products that implement it. The inference of both parameters is helpful to describe the standarization/parameterization balance of a product line, detect scope flaws, assess the product line incremental development, and improve the accuracy of economic models. The paper reports experimental evidence that our algorithm has better runtime performance than existing alternative approaches.
  • Publicación
    Exemplar driven development of software product lines
    (Elsevier, 2012-12-01) Heradio Gil, Rubén; Fernández Amoros, David José; Torre Cubillo, Luis de la; Abad Cardiel, Ismael
    The benefits of following a product line approach to develop similar software systems are well documented. Nevertheless, some case studies have revealed significant barriers to adopt such approach. In order to minimize the paradigm shift between conventional software engineering and software product line engineering, this paper presents a new development process where the products of a domain are made by analogy to an existing product. Furthermore, this paper discusses the capabilities and limitations of different techniques to implement the analogy relation and proposes a new language to overcome such limitations.
  • Publicación
    A literature review on feature diagram product counting and its usage in software product line economic models
    (World Scientific Publishing, 2013-10-01) Heradio Gil, Rubén; Fernández Amoros, David José; Cerrada Somolinos, José Antonio; Abad Cardiel, Ismael
    In software product line engineering, feature diagrams are a popular means to represent the similarities and differences within a family of related systems. In addition, feature diagrams implicitly model valuable information that can be used in economic models to estimate the cost savings of a product line. In particular, this paper reviews existing proposals on computing the total number of products modeled with a feature diagram and, given a feature, the number of products that implement it. The paper also reviews the economic information that can be estimated when such numbers are known. Thus, this paper contributes by bringing together previously-disparate streams of work: the automated analysis of feature diagrams and economic models for product lines.
  • Publicación
    Speeding up derivative configuration from product platforms
    (MDPI, 2014-06-18) Pérez Morago, Héctor José; Adán, Antonio; Heradio Gil, Rubén; Fernández Amoros, David José
    To compete in the global marketplace, manufacturers try to differentiate their products by focusing on individual customer needs. Fulfilling this goal requires that companies shift from mass production to mass customization. Under this approach, a generic architecture, named product platform, is designed to support the derivation of customized products through a configuration process that determines which components the product comprises. When a customer configures a derivative, typically not every combination of available components is valid. To guarantee that all dependencies and incompatibilities among the derivative constituent components are satisfied, automated configurators are used. Flexible product platforms provide a big number of interrelated components, and so, the configuration of all, but trivial, derivatives involves considerable effort to select which components the derivative should include. Our approach alleviates that effort by speeding up the derivative configuration using a heuristic based on the information theory concept of entropy.
  • Publicación
    Supporting commonality-based analysis of software product lines
    (Institution of Engineering and Technology (IET), 2011-03-24) Heradio Gil, Rubén; Fernández Amoros, David José; Cerrada Somolinos, José Antonio; Cerrada Somolinos, Carlos
    Software Product Line (SPL) engineering is a cost effective approach to developing families of similar products. Key to the success of this approach is to correctly scope the domain of the SPL, identifying the common and variable features of the products and the interdependencies between features. In this paper, we show how the commonality of a feature (i.e., the reuse ratio of the feature among the products) can be used to detect scope flaws in the early stages of development. SPL domains are usually modeled by means of feature diagrams following the FODA notation. We extend classical FODA trees with unrestricted cardinalities, and present an algorithm to compute the number of products modeled by a feature diagram and the commonality of the features. Finally, we compare the performance of our algorithm with two other approaches built on top of boolean logic SAT-solver technology such as cachet and relsat.
  • Publicación
    Scalable Sampling of Highly-Configurable Systems: Generating Random Instances of the Linux Kernel
    (Association for Computing Machinery (ACM), 2023-01-05) Mayr Dorn, Christoph; Egyed, Alexander; Fernández Amoros, David José; Heradio Gil, Rubén
    Software systems are becoming increasingly configurable. A paradigmatic example is the Linux kernel, which can be adjusted for a tremendous variety of hardware devices, from mobile phones to supercomputers, thanks to the thousands of configurable features it supports. In principle, many relevant problems on configurable systems, such as completing a partial configuration to get the system instance that consumes the least energy or optimizes any other quality attribute, could be solved through exhaustive analysis of all configurations. However, configuration spaces are typically colossal and cannot be entirely computed in practice. Alternatively, configuration samples can be analyzed to approximate the answers. Generating those samples is not trivial since features usually have inter-dependencies that constrain the configuration space. Therefore, getting a single valid configuration by chance is extremely unlikely. As a result, advanced samplers are being proposed to generate random samples at a reasonable computational cost. However, to date, no sampler can deal with highly configurable complex systems, such as the Linux kernel. This paper proposes a new sampler that does scale for those systems, based on an original theoretical approach called extensible logic groups. The sampler is compared against five other approaches. Results show our tool to be the fastest and most scalable one.
  • Publicación
    Anotación semántica no supervisada
    (Universidad Nacional de Educación a Distancia (España). Escuela Técnica Superior de Ingeniería Informática. Departamento de Lenguajes y Sistemas Informáticos, 2004-11-29) Fernández Amoros, David José
    En esta tesis se trata el problema de la desambiguación del sentido de las palabras (i.e. dados un diccionario, una palabra y un contexto, decidir en qué sentido del diccionario se está usando la palabra en el contexto). Las diferentes fuentes de información utilizadas son : 1. La información de origen taxonómico basada en la relación es-un, por ejemplo, un águila es-un pájaro. 2. La información de coocurrencias. Tomando como punto de partida un corpus de casi 300 millones de palabras provinientes de libros en formato electrónico (Proyecto Gutenberg) estudiaremos pares de palabras cuyas apariciones en contextos cortos son estadísticamente dependientes. Utilizaremos varias medidas para calibrar ese grado de dependencia y emplearemos dicha información para desambiguar. 3. Información extraída de la WWW. La información de la glosas del inventario de sentidos serán complementadas con información extraída de la Web. Esta información ha sido extraída de un sistema de clasificación de documentos realizado por voluntarios (Open Directory Project) por Celina Santamaría. 4. Información proviniente de corpora bilingüe comparable. Partiendo de un corpus en inglés y otro en español se han buscado patrones sintácticos superficiales correspondientes a sintagmas nominales en ambos idiomas. A partir de este trabajo realizado por Anselmo Peñas y Fernando López Ostenero estudiaremos si es posible aprovechar las diferencias entre ambos idiomas para detectar estos sintagmas y desambiguar mediante las capacidades translingües de una base de conocimiento léxica (EuroWordNet). Se demostrará que la anotación semántica no supervisada puede lograr buenos resultados, y que hay lineas de investigación, con un importante potencial de mejora, que merecen exploradas.
  • Publicación
    A Monte Carlo tree search conceptual framework for feature model analyses
    (Elsevier, 2023-01) Horcas, José Miguel; Galindo, José A.; Benavides, David; Heradio Gil, Rubén; Fernández Amoros, David José
    Challenging domains of the future such as Smart Cities, Cloud Computing, or Industry 4.0 expose highly variable systems with colossal configuration spaces. The automated analysis of those systems’ variability has often relied on SAT solving and constraint programming. However, many of the analyses have to deal with the uncertainty introduced by the fact that undertaking an exhaustive exploration of the whole configuration space is usually intractable. In addition, not all analyses need to deal with the configuration space of the feature models, but with different search spaces where analyses are performed over the structure of the feature models, the constraints, or the implementation artifacts, instead of configurations. This paper proposes a conceptual framework that tackles various of those analyses using Monte Carlo tree search methods, which have proven to succeed in vast search spaces (e.g., game theory, scheduling tasks, security, program synthesis, etc.). Our general framework is formally described, and its flexibility to cope with a diversity of analysis problems is discussed. We provide a Python implementation of the framework that shows the feasibility of our proposal, identifying up to 11 lessons learned, and open challenges about the usage of the Monte Carlo methods in the software product line context. With this contribution, we envision that different problems can be addressed using Monte Carlo simulations and that our framework can be used to advance the state-of-the-art one step forward.
  • Publicación
    Uniform and scalable sampling of highly configurable systems
    (Springer, 2022-01-21) Galindo, José A.; Benavides, David; Batory, Don; Heradio Gil, Rubén; Fernández Amoros, David José
    Many analyses on configurable software systems are intractable when confronted with colossal and highly-constrained configuration spaces. These analyses could instead use statistical inference, where a tractable sample accurately predicts results for the entire space. To do so, the laws of statistical inference requires each member of the population to be equally likely to be included in the sample, i.e., the sampling process needs to be “uniform”. SAT-samplers have been developed to generate uniform random samples at a reasonable computational cost. However, there is a lack of experimental validation over colossal spaces to show whether the samplers indeed produce uniform samples or not. This paper (i) proposes a new sampler named BDDSampler, (ii) presents a new statistical test to verify sampler uniformity, and (iii) reports the evaluation of BDDSampler and five other state-of-the-art samplers: KUS, QuickSampler, Smarch, Spur, and Unigen2. Our experimental results show only BDDSampler satisfies both scalability and uniformity.
  • Publicación
    Supporting the Statistical Analysis of Variability Models
    (Institute of Electrical and Electronics Engineers (IEEE), 2019-08-26) Mayr Dorn, Christoph; Egyed, Alexander; Heradio Gil, Rubén; Fernández Amoros, David José
    Variability models are broadly used to specify the configurable features of highly customizable software. In practice, they can be large, defining thousands of features with their dependencies and conflicts. In such cases, visualization techniques and automated analysis support are crucial for understanding the models. This paper contributes to this line of research by presenting a novel, probabilistic foundation for statistical reasoning about variability models. Our approach not only provides a new way to visualize, describe and interpret variability models, but it also supports the improvement of additional state-of-the-art methods for software product lines; for instance, providing exact computations where only approximations were available before, and increasing the sensitivity of existing analysis operations for variability models. We demonstrate the benefits of our approach using real case studies with up to 17,365 features, and written in two different languages (KConfig and feature models).