Improving memory efficiency for processing large-scale models - Inria - Institut national de recherche en sciences et technologies du numérique Accéder directement au contenu
Communication Dans Un Congrès Année : 2014

Improving memory efficiency for processing large-scale models

Résumé

Scalability is a main obstacle for applying Model-Driven Engineering to reverse engineering, or to any other activity manipulating large models. Existing solutions to persist and query large models are currently ine cient and strongly linked to memory availability. In this paper, we propose a memory unload strategy for Neo4EMF, a persistence layer built on top of the Eclipse Modeling Framework and based on a Neo4j database backend. Our solution allows us to partially unload a model during the execution of a query by using a periodical dirty saving mechanism and transparent reloading. Our experiments show that this approach enables to query large models in a restricted amount of memory with an acceptable performance.
Fichier principal
Vignette du fichier
bigmde14_submission_6_1_.pdf (325.59 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

hal-01033188 , version 1 (22-07-2014)

Identifiants

  • HAL Id : hal-01033188 , version 1

Citer

Gwendal Daniel, Gerson Sunyé, Amine Benelallam, Massimo Tisi. Improving memory efficiency for processing large-scale models. BigMDE, University of York, Jul 2014, York, UK, United Kingdom. ⟨hal-01033188⟩
393 Consultations
680 Téléchargements

Partager

Gmail Facebook X LinkedIn More