This paper presents an anaphora analysis system that was an entry for the Dialog 2014 anaphora analysis competition. The system is based on ABBY Y Compreno linguistic technologies. For some of the tasks of this competition we used basic features of the Compreno technology, while others required building new rules and mechanisms or making adjustments to the existing ones. Below we briefly describe the mechanisms (both basic and new) that were used in our system for this competition.

The midline results of a scientific survey of the definition of requirements for the development of an innovative ECM system are presented. These results were obtained during the second stage of a complex project carried out through Government Grant with participation of NRU and IT Corporation (Russia). The particular research is based on 3 scopes: 1) Enterprise Content Management (ECM) system as roofing, 2) umbellate Total Content Management system (TCM); 3) process-based enterprise innovation activity and subject-oriented approach to BPM.

Statement of Research. A need to reduce the increasing number of system vulnerabilities caused by unauthorized software installed on computer aids necessitates development of an approach to automate the data-storage media audit. The article describes an approach to identification of informative assembly instructions. Also, the influence of a chosen feature that is used to create a unified program signature on identification result is shown. Methods. Shannon method allowing a determination of feature informativeness for a random number of object classes and not depending on the sample volume of observed features is used to calculate informativeness. Identification of elf-files was based on applying statistical chi-squared test of homogeneity. Main Findings. Quantitative characteristics of informativeness for 118 assembly instructions have been obtained. The analysis of experimental results for executable files identification with 10 different features used to create program signatures compared by means of the chi-squared test of homogeneity at significance levels p = 0.05 and p = 0.01 has been carried out. Practical Relevance. The importance of using a particular feature in program signature creation has been discovered, as well as the capability of considering several executable file signatures together to provide a summative assessment on their belonging to a certain program.

The approaches based on applying of metamodeling and domain-specific languages are widely used in software engineering. There are many different tools for creating visual domain-specific modeling languages with a possibility of determining user’s graphical notations. However, these tools possess disadvantages. The article presents an approach to the development of language workbench that allows to eliminate some restrictions of existing DSM-platforms. The MetaLanguage system is designed for creation of visual dynamic adaptable domain-specific modeling languages and for models construction with these languages. It allows executing transformations of the created models in various textual and graphical notations. Basic metalanguage constructions of this system are described. The formal description of modeling languages metamodel used in MetaLanguage is given. The architecture of MetaLanguage toolkit is presented.

The paper discusses a new approach to developing tools for quantitatively analyzing the financial behavior of small and medium price-taking traders each possessing abilities to predict share price values for a set of financial securities traded in a stock exchange. Tools for forming and managing a trader’s portfolio of securities from this set are proposed. Particularly, it is shown that when the trader can treat share price values from the portfolio as random variables with known (to her) distributions, an optimal portfolio composition is found by solving a linear programming problem. Otherwise, this optimal composition is found as the trader’s equilibrium strategy in an antagonistic two-person game with the stock exchange being the other player. In this game on polyhedra of disjoint player strategies, described by systems of linear equations and inequalities of a balance kind, calculating saddle points is reduced to solving linear programming problems forming a dual pair.

*Abstract – *Nowadays approaches, based on models, are used in the development of the information systems. The models can be changed during the system development process by developers. They can be transformed automatically: visual model can be translated into program code; transformation from one modeling language to other can be done. The most appropriate way of the formal visual model presentation is metagraph. The best way to describe changes of visual models is the approach, based on graph grammars (graph rewriting). It is the most demonstrative way to present the transformation. But applying the graph grammar to the graph of model means to find the subgraph isomorphic to the left part of the grammar rule. This is an NP-complete task. There are some algorithms, developed for solving this task. They were designed for ordinary graphs and hypergraphs. In this article we consider some of them in case of using with the metagraphs representing models.

The paper proposes an approach to instruction stream generation for verification of microprocessor designs. The approach is based on using formal specifications of the instruction set architecture as a source of knowledge about the design under verification. This knowledge is processed with generic engines implementing an extensible set of generation strategies to produce stimuli in the form of instruction sequences. Generation tasks are formulated using high-level descriptions that specify target instructions and strategies of sequence construction and data generation. This provides a flexible way to generate deterministic, random and constraint-based stimuli for verification of arbitrary architectures with minimum effort. The proposed approach has been successfully applied in industrial projects for verification of ARMv8 and MIPS64 microprocessor designs.

Statement of Research. A need to reduce the increasing number of system vulnerabilities caused by unauthorized software installed on computer aids necessitates development of an approach to automate the data-storage media audit. The article describes an approach to identification of informative assembly instructions. Also, the influence of a chosen feature that is used to create a unified program signature on identification result is shown. Methods. Shannon method allowing a determination of feature informativeness for a random number of object classes and not depending on the sample volume of observed features is used to calculate informativeness. Identification of elf-files was based on applying statistical chi-squared test of homogeneity. Main Findings. Quantitative characteristics of informativeness for 118 assembly instructions have been obtained. The analysis of experimental results for executable files identification with 10 different features used to create program signatures compared by means of the chi-squared test of homogeneity at significance levels p = 0.05 and p = 0.01 has been carried out. Practical Relevance. The importance of using a particular feature in program signature creation has been discovered, as well as the capability of considering several executable file signatures together to provide a summative assessment on their belonging to a certain program.

Existing diversity of types of formal cognitive maps with corresponding theoretical models makes actual the problem of their systematization for further comparative analysis of their capabilities in solving applied problems and for development of the general theory of formal cognitive maps, when and if it is possible. The lack of the general theory of formal cognitive maps, orientation of many known theoretical models exclusively on simulation, semantic vagueness of some models have already led to incorrect formal statements and solutions of applied problems on the basis of cognitive maps. The general parametrical model of semantics of functional cognitive maps is developed to uniformly describe semantics of formal cognitive maps. The model has covered the major part of known types of formal cognitive maps and has enabled systemizing non-functional types of maps. Efficiency of the proposed approach to systematization of formal cognitive maps is shown on the representative set of types of such maps.

An approach to multiple labeling research papers is explored. We develop techniques for annotating/labeling research pa- pers in informatics and computer sciences with key phrases taken from the ACM Computing Classification System. The techniques utilize a phrase-to-text relevance measure so that only those phrases that are most relevant go to the anno- tation. Three phrase-to-text relevance measures are experi- mentally compared in this setting. The measures are: (a) co- sine relevance score between conventional vector space repre- sentations of the texts coded with tf-idf weighting; (b) pop- ular characteristic of probability of term generation BM25; and (c) an in-house characteristic of conditional probability of symbols averaged over matching fragments in suffix trees representing texts and phrases, CPAMF. In an experiment conducted over a set of texts published in journals of the ACM and manually annotated by their authors, CPAMF outperforms both the cosine measure and BM25 by a wide margin.

Today many problems that are dedicated to a particular problem domain can be solved using DSL. Thus to use DSL it must be created or it can be selected from existing ones. Creating a completely new DSL in most cases requires high financial and time costs. Selecting an appropriate existing DSL is an intensive task because such actions like walking through every DSL and deciding if current DSL can handle the problem are done manually. This problem appears because there are no DSL repository and no tools for matching suitable DSL with specific task. This paper observes an approach for implementing an automated detection of requirements for DSL (ontology-based structure) and automated DSL matching for specific task.

We consider the scheduling problems α|β|Fmax: A set of n jobs J1,…,Jn with release dates r1,…,rn, processing times p1,…,pn and due dates d1,…,dn has to be scheduled on a single or many machines. The job preemption is not allowed. The goal is to find aschedule that minimizes the regular function F(C1,…,Cn), that Cj is the job jcompletion time. We have suggest an approximation scheme to find approximate optimal value of the objective function.

In this paper, for $NP$-hardness single and multi-machine scheduling problems with the criterion of minimization maximum lateness the metrics $\rho$ has been used. We consider some approaches finding of the approximate solution for the problems. The idea of approaches consists in construction to a initial instance $A$ such instance $B$ (with the same number of jobs) with minimum of estimation of absolute error that$$0\le L_{max}^A(\pi^B)-L_{max}^A(\pi^A)\le \rho_d(A,B)+ \rho_r(A,B)+\rho_p(A,B),$$ where $\rho_d(A,B)=\max\limits_{j\in N}\{d_j^A-d_j^B\}-\min\limits_{j\in N}\{d_j^A-d_j^B\},$ $ \rho_r(A,B)=\max\limits_{j\in N}\{r_j^A-r_j^B\}-\min\limits_{j\in N}\{r_j^A-r_j^B\}$ and$\rho_p(A,B)=\sum\limits_{j\in N}|p_j^A-p_j^B|,$ and $\pi^A, \pi^B$ -- optimal schedules for instances $A$ and $B$,respectively. Besides $\rho(A,B)=\rho_d(A,B)+ \rho_r(A,B)+\rho_p(A,B)$ satisfies to properties of the metrics in$(3n-2)$-dimensional space $\{(r_j,p_j,d_j)\,|\,j\in N\}$ with fixed in two parameters.

In the paper, we consider the NP-hard total tardiness minimization on a single machine scheduling problem. We propose a metric for that problem and present a new polynomial approximation scheme based on search for the polynomially solvable instance which has a minimal distance in the metric from an initial instance.

An arithmetic theory of oppositions is devised by comparing expressions, Boolean bitstrings, and integers. This leads to a set of correspondences between three domains of investigation, namely: logic, geometry, and arithmetic. The structural properties of each area are investigated in turn, before justifying the procedure as a whole. To finish, I show how this helps to improve the logical calculus of oppositions, through the consideration of corresponding operations between integers.

We consider a model of regions’ ranking in terms of their vulnerability to natural and technological disasters. Regions are different in terms of their resistance to different disasters, by their population, by the distribution of the sources of potential disasters, etc. We consider different models of a data envelopment analysis (DEA) approach taking into account the risks of the implementation of different measures, their cost as well as the heterogeneity of regions. The numerical examples demonstrate the application of the constructed model for the regions of Russian Federation.

The World Trade Organization ( WTO) accession opened new possibilities for Russian economy and imposed new restrictions on economic policies. Three years after the accession Russia still is in transition period. The main questions posed in this paper are: what the net effect will be after the transition period ends and what can be expected in specific sectors and regions due to the changes that will take place.

Abstract. A suffix-tree based method for measuring similarity of a key phrase to an unstructured text is proposed. The measure involves less computation and it does not depend on the length of the text or the key phrase. This applies to the following tasks in semantic text analysis:

Finding interrelations between key phrases over a set of texts;

Annotating a research article by topics from a taxonomy of the domain;

Clustering relevant topics and mapping clusters on a domain taxonomy.

Ideas and achievements of Anatoly Kitov in the development of nationwide and local industrial management systems are described and analyzed in this report in broad economic and political context.

*Šimâ* *milka* is an Akkadian literary text belonging to the genre of instructions. It survives in manuscripts from Ugarit, Emar, Hattuša and Kalhu. This article is the ﬁrst of a series in which a complete new edition of this composition will hopefully be published.