scispace - formally typeset
Search or ask a question

Showing papers on "Serialization published in 2017"


Journal ArticleDOI
TL;DR: This approach combines an application programing interface (API) inspired by pandas with the Common Data Model for self-described scientific data to provide a toolkit and data structures for N-dimensional labeled arrays.
Abstract: xarray is an open source project and Python package that provides a toolkit and data structures for N-dimensional labeled arrays. Our approach combines an application programing interface (API) inspired by pandas with the Common Data Model for self-described scientific data. Key features of the xarray package include label-based indexing and arithmetic, interoperability with the core scientific Python packages (e.g., pandas, NumPy, Matplotlib), out-of-core computation on datasets that don’t fit into memory, a wide range of serialization and input/output (I/O) options, and advanced multi-dimensional data manipulation tools such as group-by and resampling. xarray, as a data model and analytics toolkit, has been widely adopted in the geoscience community but is also used more broadly for multi-dimensional data analysis in physics, machine learning and finance.

620 citations


01 Jul 2017
TL;DR: This document defines the CBOR Object Signing and Encryption (COSE) protocol, which describes how to create and process signatures, message authentication codes, and encryption using CBOR for serialization.
Abstract: Concise Binary Object Representation (CBOR) is a data format designed for small code size and small message size. There is a need for the ability to have basic security services defined for this data format. This document defines the CBOR Object Signing and Encryption (COSE) protocol. This specification describes how to create and process signatures, message authentication codes, and encryption using CBOR for serialization. This specification additionally describes how to represent cryptographic keys using CBOR.

83 citations


Journal ArticleDOI
TL;DR: The analysis of results indicates that ifcJSON4 schema developed in this paper is a valid JSON schema that can guide the creation of valid ifc JSON documents to be used for web-based data transfer and to improve interoperability of Cloud-based BIM applications.

77 citations


Journal ArticleDOI
TL;DR: An alternative systematic methodology for parameter tuning is proposed, which can be easily applied onto any computing infrastructure and is shown to yield comparable if not better results than the initial one when applied to MN3; observed speedups in the validating test case studies start from 20%.

56 citations


Journal ArticleDOI
TL;DR: The authors advocate the use of binary storage for sizeable point cloud scans, but also show how especially the grid discretization can result into usable points cloud segments embedded into text-based IFC models.

36 citations


Journal ArticleDOI
01 Jun 2017
TL;DR: This paper explores and analyse the result set serialization design space, presents experimental results from a large chunk of the database market, and proposes a columnar serialization method that improves transmission performance by an order of magnitude.
Abstract: Transferring a large amount of data from a database to a client program is a surprisingly expensive operation. The time this requires can easily dominate the query execution time for large result sets. This represents a significant hurdle for external data analysis, for example when using statistical software. In this paper, we explore and analyse the result set serialization design space. We present experimental results from a large chunk of the database market and show the inefficiencies of current approaches. We then propose a columnar serialization method that improves transmission performance by an order of magnitude.

28 citations


Journal ArticleDOI
TL;DR: This paper presents MpOS (Multiplatform Offloading System), a framework that supports a method-based offloading technique for applications of different mobile platforms (Android and Windows Phone) and shows that the type of serialization used by the framework directly impacts on the offloading performance.

25 citations


Journal ArticleDOI
01 Aug 2017
TL;DR: Experiments show that applications using Fad.js achieve speedups up to 2.7x for encoding and 9.9x for decoding JSON data when compared to state-of-the art JSON processing libraries.
Abstract: JSON is one of the most popular data encoding formats, with wide adoption in Databases and BigData frameworks as well as native support in popular programming languages such as JavaScript/Node.js, Python, and R.Nevertheless, JSON data processing can easily become a performance bottleneck in data-intensive applications because of parse and serialization overhead. In this paper, we introduce Fad.js, a runtime system for efficient processing of JSON objects in data-intensive applications. Fad.js is based on (1) speculative just-in-time (JIT) compilation and (2) selective access to data. Experiments show that applications using Fad.js achieve speedups up to 2.7x for encoding and 9.9x for decoding JSON data when compared to state-of-the art JSON processing libraries.

22 citations


Journal ArticleDOI
TL;DR: The article discusses how the micro-serialization of Twitter fiction both differs from and draws on the pre-digital tradition of serial fiction, and focuses on two interrelated aspects of serialization, temporality and interaction.
Abstract: The final part of the recent anthology Serialization in Popular Culture (2014) is called ‘Digital serialization’ and is devoted to ‘the influence of digital technologies on serial form’. The chapte...

15 citations


Proceedings ArticleDOI
01 Jul 2017
TL;DR: The paper shows that there are better alternatives than XML & JAXB and gives guidance in choosing the most appropriate serialization format and library depending on the context, especially in the context of the Internet of Things.
Abstract: Communication between DERs and System Operators is required to provide Demand Response and solve some of the problems caused by the intermittency of much Renewable Energy. An important part of efficient communication is serialization, which is important to ensure a high probability of delivery within a given timeframe, especially in the context of the Internet of Things, using low-bandwidth data connections and constrained devices. The paper shows that there are better alternatives than XML & JAXB and gives guidance in choosing the most appropriate serialization format and library depending on the context.

14 citations


Journal ArticleDOI
TL;DR: In this paper, a semantic annotation of services is supported through ontologies defined for API definition languages such as Swagger and RAML, which enables intelligent discovery of services in IoT environments.

Proceedings ArticleDOI
01 Dec 2017
TL;DR: This study aims to clarify efficient/inefficient use of disks in intermediate data caching in Spark and also to improve the usability of disks for end users.
Abstract: Apache Spark is a parallel data processing framework that executes fast for iterative calculations and interactive processing, by caching intermediate data in memory with a lineage-based data recovery from faults. The Spark system can also manage data sets larger than memory capacity by placing some cache or all of them on disks on processing nodes. However, the disadvantage is potential performance degradation due to disk I/O and/or serialization. This study aims to clarify efficient/inefficient use of disks in intermediate data caching in Spark and also to improve the usability of disks for end users. In order to achieve the purpose, influence of disk use in data caching was firstly investigated in various aspects, such as caching options, data abstractions and storage devices. The results indicate that serialization cost is dominant rather than disk I/O in most cases. Secondly, a method of combined use of memory and disk was further evaluated under a high memory pressure. Then the method was improved to avoid an excessive re-caching problem, which achieved at most 20–30% reduction of total execution time under a high memory pressure and did not degrade the performance under a low memory pressure, in our experiment with 4 machine learning benchmarks. Finally, this paper summarizes important factors and potential improvements for efficiently using disks in data caching in Spark.

Patent
27 Jun 2017
TL;DR: In this paper, the authors proposed a collaborative authentication platform and processes, combining digital fingerprint databases with associated incentive databases, to provide unparalleled reliability and enriched metadata to supply chain tracking, detecting counterfeit objects, and other applications.
Abstract: Databases storing digital fingerprints of physical objects enable enhanced security and collaborative authentication. Digital fingerprints enable reliable identification of an object without the need for attaching or associating physical tags, labels or other identifying materials with the object; and serialization for identification also is obviated. By combining digital fingerprinting and data collaboration in one process, parties to the data collaboration can gain a level of certainty that data attributed to an object by different parties or at different times is attributed to only that object and not erroneously attributed to an incorrect or counterfeit object. Collaborative authentication platforms and processes, combining digital fingerprint databases with associated incentive databases, contribute enhanced information to the authentication databases, and provide unparalleled reliability and enriched metadata to supply chain tracking, detecting counterfeit objects, and other applications.


Proceedings ArticleDOI
14 Jun 2017
TL;DR: It is claimed that PhTM is a competitive alternative to HyTM and proposed PhTM*, the first implementation of PhTM on modern HTM-ready processors, which relies in avoiding unnecessary transitions to software mode by taking into account the categories of hardware aborts and adding a new serialization mode.
Abstract: In recent years, Hybrid TM (HyTM) has been proposed as a transactional memory approach that leverages on the advantages of both hardware (HTM) and software (STM) execution modes. HyTM assumes that concurrent transactions can have very different phases and thus should run under different execution modes. Although HyTM has shown to improve performance, the overall solution can be complicated to manage, both in terms of correctness and performance. On the other hand, Phased Transactional Memory (PhTM) considers that concurrent transactions have similar phases, and thus all transactions could run under the same mode. As a result, PhTM does not require coordination between transactions on distinct modes making its implementation simpler and more flexible. In this paper we claim that PhTM is a competitive alternative to HyTM and propose PhTM*, the first implementation of PhTM on modern HTM-ready processors. PhTM* novelty relies in avoiding unnecessary transitions to software mode by: (i) taking into account the categories of hardware aborts; (ii) adding a new serialization mode. Experimental results with Haswell's TSX reveal that, for the STAMP benchmark suite, PhTM* performs on average 11% better than PhTM, a previous phase-based TM, and 15% better than HyTM-NOrec, a state-of-the-art HyTM. In addition, PhTM* showed to be even more effective running on a Power8 machine by performing over 25% and 36% better than PhTM and HyTM-NOrec, respectively.

Book ChapterDOI
28 May 2017
TL;DR: A visual representation based on a block metaphor for creating and editing such mappings that is fully compliant with the R2RML specification is described and preliminary findings from users using the tool indicate that the visual representation was helpful in the creation of R1RML mappings with good usability results.
Abstract: R2RML is the W3C standard mapping language used to define customized mappings from relational databases into RDF. One issue that hampers its adoption is the effort needed in the creation of such mappings, as they are stored as RDF documents. To address this problem, several tools that represent mappings as graphs have been proposed in the literature. In this paper, we describe a visual representation based on a block metaphor for creating and editing such mappings that is fully compliant with the R2RML specification. Preliminary findings from users using the tool indicate that the visual representation was helpful in the creation of R2RML mappings with good usability results. In future work, we intend to conduct more experiments focusing on different types of users and to abstract the visual representation from the R2RML mapping language so that it supports the serialization of other uplift mapping languages.

Proceedings ArticleDOI
14 Jun 2017
TL;DR: This paper builds on the underlying theory of symbolic transducers to fuse pipelines of effectful comprehensions into a single representation, from which efficient code can be generated and can significantly reduce the complexity of the fused pipelines.
Abstract: List comprehensions provide a powerful abstraction mechanism for expressing computations over ordered collections of data declaratively without having to use explicit iteration constructs. This paper puts forth effectful comprehensions as an elegant way to describe list comprehensions that incorporate loop-carried state. This is motivated by operations such as compression/decompression and serialization/deserialization that are common in log/data processing pipelines and require loop-carried state when processing an input stream of data. We build on the underlying theory of symbolic transducers to fuse pipelines of effectful comprehensions into a single representation, from which efficient code can be generated. Using background theory reasoning with an SMT solver, our fusion and subsequent reachability based branch elimination algorithms can significantly reduce the complexity of the fused pipelines. Our implementation shows significant speedups over reasonable hand-written code (3.4×, on average) and traditionally fused version of the pipeline (2.6×, on average) for a variety of examples, including scenarios for extracting fields with regular expressions, processing XML with XPath, and running queries over encoded data.

Journal ArticleDOI
TL;DR: jsCoq as discussed by the authors is a new platform and user environment for the Coq interactive proof assistant, which targets the HTML5-ECMAScript 2015 specification, and it is typically run inside a standards-compliant browser, without the need of external servers or services.
Abstract: We describe jsCcoq, a new platform and user environment for the Coq interactive proof assistant. The jsCoq system targets the HTML5-ECMAScript 2015 specification, and it is typically run inside a standards-compliant browser, without the need of external servers or services. Targeting educational use, jsCoq allows the user to start interaction with proof scripts right away, thanks to its self-contained nature. Indeed, a full Coq environment is packed along the proof scripts, easing distribution and installation. Starting to use jsCoq is as easy as clicking on a link. The current release ships more than 10 popular Coq libraries, and supports popular books such as Software Foundations or Certified Programming with Dependent Types. The new target platform has opened up new interaction and display possibilities. It has also fostered the development of some new Coq-related technology. In particular, we have implemented a new serialization-based protocol for interaction with the proof assistant, as well as a new package format for library distribution.

Patent
08 Dec 2017
TL;DR: In this article, a CMOS image data training system and an image data serialization-deserialization simulation detection method, relates to the CMOS data serialisation-deerialization simulations detection method and solves the problem that the data serializatio-de-erialization is difficult to carry out due to the fact that an uncertain phase relationship exists among transmission channels employed by a CIMOS image sensor each time when electrification is carried out.
Abstract: The invention discloses a CMOS image data training system and an image data serialization-deserialization simulation detection method, relates to the CMOS image data serialization-deserialization simulation detection method and solves the problem that the data serialization-deserialization is difficult to carry out due to the fact that an uncertain phase relationship exists among transmission channels employed by a CMOS image sensor each time when electrification is carried out. The CMOS image data training system comprises the CMOS image sensor and a data processor. The interior of the data processor is composed of iodelay, an iserdes, data asynchronous FIFO, control asynchronous FIFO, a gearbox, a ram based shifter and a controller. The controller is taken as the core of the CMOS image data training system and is used for controlling various components to work coordinatively. Under the control of the controller, the CMOS image sensor outputs serial graph data and finally the serial graph data is converted into parallel image data with bit width of p through the iodelay, the iserdes, the data asynchronous FIFO, the control asynchronous FIFO, the gearbox 1:2 and the ram based shifter. According to the serialization-deserialization detection method based on simulation provided by the invention, different excitations are generated for different data training phases, thereby realizing different training strategies.

Patent
31 May 2017
TL;DR: In this article, the authors proposed a data serialization method and device, which comprises the steps that original data is acquired; for each field name in a preset data template, field content corresponding the field name is acquired from the original data, wherein an association relationship among the field names, data types and field serial numbers is pre-stored in the preset data templates, and the data types are data types of the field content.
Abstract: The invention provides a data serialization method and device. The method comprises the steps that original data is acquired; for each field name in a preset data template, field content corresponding the field name is acquired from the original data, wherein an association relationship among the field names, data types and field serial numbers is pre-stored in the preset data template, and the data types are data types of the field content; the field content corresponding to the field names and the field serial numbers associated to the field names are serialized according to the serialization modes corresponding to the data types associated to the field names, and byte sequences corresponding to fields to which the field names belong are obtained; according to the from small to big order of the field serial numbers, the obtained byte sequences corresponding to the fields to which the field names belong are assembled into one byte sequence to serve as serialized data corresponding to the original data. According to the data serialization method and device, the data occupation space can be effectively decreased, the data transmission and disk read-write efficiency can be improved.

Proceedings ArticleDOI
25 May 2017
TL;DR: The class of "calc-regular languages" is proposed, a minimalistic extension of regular languages with the additional property of handling length-fields, which disproves the conjecture that parsing those languages is difficult and not efficient.
Abstract: When binary data are sent over a byte stream, the binary format sender and receiver are using is a "data serialization language", either explicitely specified, or implied by the implementations. Security is at risk when sender and receiver disagree on details of this language. If, e.g., the receiver fails to reject invalid messages, an adversary may assemble such invalid messages to compromise the receiver's security. Many data serialization languages are length-prefix languages. When sending/storing some F of flexible size, F is encoded at the binary level as a pair (|F|, F), with |F| representing the length of F (typically in bytes). This paper's main contributions and results are as follows. (1) Length-prefix langages are not context-free. This might seem to justify the conjecture that parsing those languages is difficult and not efficient. (2) The class of "calc-regular languages" is proposed, a minimalistic extension of regular languages with the additional property of handling length-fields. Calc-regular languages can be specified via "calc-regular expressions", a natural extension of regular expressions. (3) Calc-regular languages are almost as easy to parse as regular languages, using finite-state machines with additional accumulators. This disproves the conjecture from (1).

13 Mar 2017
TL;DR: A simple, reusable data model is described with a representation format, using a well known set of keywords to expose hypermedia controls, which inform clients how to perform state transfer operations on resources.
Abstract: The scale and scope of the worldwide web has been in part driven by the availability of HTML as a common serialization, data model, and interaction model for structured resources on the web. By contrast, the general use of appropriate hypermedia techniques for machine interfaces has been limited by the lack of a common format for serialization and exchange of structured machine resources and sensor/actuator data which includes or embeds standardized hypermedia controls. The IRTF Thing to Thing Research Group [T2TRG] has a charter to investigate the use of REST design style [REST]for machine interactions. The W3C Web of Things Interest Group [W3C-WoT] are investigating abstract hypermedia controls and interaction models for machines. Machine optimized content formats exist for web links [RFC5988] [RFC6690] and for data items [I-D.ietf-core-senml]. Structured data which contains both links and items is known as the collection pattern. This draft describes media types for representation of machine resources structured as collections. A simple, reusable data model is described with a representation format, using a well known set of keywords to expose hypermedia controls, which inform clients how to perform state transfer operations on resources. The underlying assumptions regarding transfer layer processing are specified in this document. The HSML media type described in this document is compatible with SenML and CoRE Link- format by reusing the keyword identifiers and element structures from these content formats. Representations of HSML document content may be obtained in CoRE Link-Format and SenML content formats.

Journal ArticleDOI
20 Nov 2017
TL;DR: A new mechanism to persistify is proposed, taking an object which lives in memory only and storing it on disk as a persistent object and serve the geometry of HEP experiments, and a new generation of applications could be developed, which can use the actual detector geometry while being platform-independent and experiment-independent.
Abstract: The complex geometry of the whole detector of the ATLAS experiment at LHC is currently stored only in custom online databases, from which it is built on-the-fly on request. Accessing the online geometry guarantees accessing the latest version of the detector description, but requires the setup of the full ATLAS software framework "Athena", which provides the online services and the tools to retrieve the data from the database. This operation is cumbersome and slows down the applications that need to access the geometry. Moreover, all applications that need to access the detector geometry need to be built and run on the same platform as the ATLAS framework, preventing the usage of the actual detector geometry in stand-alone applications. Here we propose a new mechanism to persistify (in software development in general, and in HEP computing in particular, persistifying means taking an object which lives in memory only - for example because it was built on-the-fly while processing the experimental data, - serializing it and storing it on disk as a persistent object) and serve the geometry of HEP experiments. The new mechanism is composed by a new file format and the modules to make use of it. The new file format allows to store the whole detector description locally in a file, and it is especially optimized to describe large complex detectors with the minimum file size, making use of shared instances and storing compressed representations of geometry transformations. Then, the detector description can be read back in, to fully restore the in-memory geometry tree. Moreover, a dedicated REST API is being designed and developed to serve the geometry in standard exchange formats like JSON, to let users and applications download specific partial geometry information. With this new geometry persistification a new generation of applications could be developed, which can use the actual detector geometry while being platform-independent and experiment-independent.

Proceedings ArticleDOI
01 Sep 2017
TL;DR: There are better alternatives to using Web Services and XMPP as middleware and that there is better alternatives than using XML for serialization and ZeroMQ, YAMI4, and ICE are the middleware that performs the best, and ProtoBuf (ProtoStuff), and ProtoStuff are the serialization that performsThe best.
Abstract: To solve the problems caused by intermittent renewable energy production, communication between Distributed Energy Resources (DERs) and system operators is necessary. The communication middleware and serialization used for communication are essential to ensure delivery of the messages within the required timeframe, to provide the necessary ancillary services to the power grid. This paper shows that there are better alternatives to using Web Services and XMPP as middleware and that there are better alternatives than using XML for serialization. The paper also gives guidance at choosing the best communication middleware and serialization format/library, aided by the authors' earlier work, which investigates the performance and characteristics of communication middleware and serialization independently. Given the performance criteria of the paper, ZeroMQ, YAMI4, and ICE are the middleware that performs the best, and ProtoBuf (ProtoStuff), and ProtoStuff are the serialization that performs the best.

Patent
19 Apr 2017
TL;DR: In this article, a data object serialization method for data stream deserialization is described, which comprises the steps of obtaining a metadata description file of a to-be-serialized data object, obtaining corresponding attribute operation classes according to attribute operation class names provided by attribute description in the metadata description, and reading attribute values in the to be-serialised data object according to reading methods provided by the corresponding attributed operation classes.
Abstract: The invention discloses a data object serialization method and apparatus, a data stream deserialization method and apparatus, an electronic device, and a serialization and deserialization system. The data object serialization method comprises the steps of obtaining a metadata description file of a to-be-serialized data object; obtaining corresponding attribute operation classes according to attribute operation class names provided by attribute description in the metadata description file; reading attribute values in the to-be-serialized data object according to reading methods provided by the corresponding attribute operation classes; and writing the read attribute values in a result data stream of the data object according to a sequence of attribute description serial numbers. By adopting the method, the problem in compatibility between systems after change of object attributes is solved; the metadata of the object and the values of the object are split; the storage space and the transmission flow are saved during storage and transmission; and the processing overhead of the systems is reduced during deserialization.

Proceedings ArticleDOI
24 Sep 2017
TL;DR: The question is asked is: How significant is the performance hit associated with choosing a particular physical implementation?
Abstract: Many cloud-based data management and analytics systems support complex objects. Dataflow platforms such as Spark and Flink allow programmers to manipulate sets consisting of objects from a host programming language (often Java). Document databases such as MongoDB make use of hierarchical interchange formats---most popularly JSON---which embody a data model where individual records can themselves contain sets of records. Systems such as Dremel and AsterixDB allow complex nesting of data structures. Clearly, no system designer would expect a system that stores JSON objects as text to perform at the same level as a system based upon a custom-built physical data model. The question we ask is: How significant is the performance hit associated with choosing a particular physical implementation? Is the choice going to result in a negligible performance cost, or one that is debilitating? Unfortunately, there does not exist a scientific study of the effect of physical complex model implementation on system performance in the literature. Hence it is difficult for a system designer to fully understand performance implications of such choices. This paper is an attempt to remedy that.

Journal ArticleDOI
TL;DR: In this paper, the notions of performativity and performance in digital environments from the combined perspective of linguistic anthropology and folkloristics are explored, and an intermediary heuristic term of "performative enactments" is introduced.
Abstract: This article explores the notions of performativity and performance in digital environments from the combined perspective of linguistic anthropology and folkloristics. In order to bring these diverging conceptual, methodological, and disciplinary traditions into mutual contact, an intermediary heuristic term of “performative enactments” is introduced. Performative enactments are elaborated as events of communicative sign behavior that foreground and make use of the principle of performativity, although not performances proper in the sense of manifesting a specific “mode of communication” (Bauman 1984). Two different cases of digital communication are analyzed, the first manifesting an instance of everyday SMS messaging between two friends, the second concerning the so-called Per-Looks media event that took place in Finland in October 2012. Both cases are approached as materially durable performative enactments with methodological attention laid on poetic patterning understood as a textually diffuse form of performativity.

26 Jul 2017
TL;DR: It is demonstrated that D2Refine is a useful and promising platform that would help address the emergent needs for clinical research study data element harmonization and standardization.
Abstract: In this paper, we present a platform known as D2Refine for facilitating clinical research study data element harmonization and standardization D2Refine is developed on top of OpenRefine (formerly Google Refine) and leverages simple interface and extensible architecture of OpenRefine D2Refine empowers the tabular representation of clinical research study data element definitions by allowing it to be easily organized and standardized using reconciliation services D2Refine builds on valuable built-in data transformation features of OpenRefine to bring source data sets to a finer state quickly We implemented the reconciliation services and search capabilities based on the standard Common Terminology Services 2 (CTS2) and the serialization of clinical research study data element definitions into standard representation using clinical information modeling technology for semantic interoperability We demonstrate that D2Refine is a useful and promising platform that would help address the emergent needs for clinical research study data element harmonization and standardization

Patent
12 Dec 2017
TL;DR: In this paper, a serialization and deserialization method for data serialization is described, in which the serialization instruction comprises object type information, according to which the object type is extracted from the data, and according to the data type, a serialisation object corresponding to the type information is created.
Abstract: The invention discloses serialization and deserialization methods and apparatuses, a computer device and a storage medium. The data serialization method in one embodiment comprises the steps of receiving a serialization instruction, wherein the serialization instruction comprises object type information; according to the object type information, calling a serialization and deserialization interface function, and creating a serialization object corresponding to the object type information; adding a serialization object instance of the serialization object; and performing data serialization by the serialization object instance, and outputting a serialization file. According to the scheme, the serialization and deserialization processing processes are simply and conveniently realized; the code quantity is reduced; and the error probability is reduced.

Proceedings ArticleDOI
01 Jul 2017
TL;DR: DCL-KDM is an efficient alternative to to generate instances of the Structure metamodel as a PA and to serialize it, and a strategy for the serialization of PAs as a Structure metAModel instance without modifying it is proposed.
Abstract: Architecture-Driven Modernization (ADM) intends to standardize software reengineering by relying on a family of standard metamodels. Knowledge-Discovery Metamodel (KDM) is the main ADM ISO metamodel aiming at representing all aspects of existing legacy systems. One of the internal KDM metamodels is called Structure, responsible for representing architectural abstractions (Layers, Components and Subsystems) and their relationships. Planned Architecture (PA) is an artifact that involves not only the architectural abstractions of the system but also the access rules that must exist between them and be maintained over time. Although PAs are frequently used in Architecture-Conformance Checking processes, up to this moment, there is no contribution showing how to specify and serialize PAs in ADM-based modernization projects. Therefore, in this paper we present an approach that i) involves a DSL (Domain-Specific Language) for the specification of PAs using the Structure metamodel concepts, and ii) proposes a strategy for the serialization of PAs as a Structure metamodel instance without modifying it. We have conducted a comparison between DCL-KDM and other techniques for specifying and generating PAs. The results showed that DCL-KDM is an efficient alternative to to generate instances of the Structure metamodel as a PA and to serialize it.