Bauman Moscow State Technical University
Author list of organization
List of articles written by the authors of the organization
-
SYSTEM OF MODELS AND METHODS OF TESTING THE HOMOGENEITY OF TWO INDEPENDENT SAMPLES
08.00.13 Mathematical and instrumental methods of Economics
DescriptionThe new paradigm of mathematical research methods allows us to give a systematic analysis of various statements of statistical analysis problems and methods for solving them, based on a probabilistic-statistical model of generating data accepted by the researcher. Methods for testing the homogeneity of two independent samples - a classic area of mathematical statistics. For more than 110 years since the publication of the fundamental Student’s article, various criteria have been developed for testing the statistical hypothesis of homogeneity in various statements, and their properties have been studied. However, the need for streamlining the totality of the scientific results found is urgent. It is necessary to analyze the whole variety of problem statements for testing the statistical hypotheses of the homogeneity of two independent samples, as well as the corresponding statistical criteria. This analysis is devoted to this article. It contains a summary of the main results concerning the methods for testing the homogeneity of two independent samples, and a comparative study of them, allowing the system to analyze the diversity of such methods in order to select the most appropriate for processing specific data. Based on the basic probabilistic-statistical model, the main statements of the problem of testing the homogeneity of two independent samples are formulated. A comparative analysis of the Student and Cramer - Welch criteria, designed to test the homogeneity of mathematical expectations, is given, a recommendation on the widespread use of the Cramer - Welch criterion is substantiated. From nonparametric methods for testing homogeneity, the criteria of Wilcoxon, Smirnov, Lehmann - Rosenblatt are considered. Dismantled two myths about the Wilcoxon criteria. Based on the analysis of the publications of the founders, the incorrectness of the term "Kolmogorov – Smirnov criterion" is shown. To verify absolute homogeneity, i.e. coincidence of the distribution functions of samples, it is recommended to use the Lehmann - Rosenblatt criterion. The current problems of the development and application of nonparametric criteria are discussed, including the difference between nominal and real significance levels, making it difficult to compare power of criteria, and the need to take into account coincidences of sample values (from the point of view of the classical theory of mathematical statistics, the probability of coincidences is 0)
-
EXISTENCE OF ASYMPTOTICALLY OPTIMAL PLANS IN DISCRETE PROBLEMS OF DYNAMIC PROGRAMMING
08.00.13 Mathematical and instrumental methods of Economics
DescriptionDynamic programming is designed to solve discrete optimal control problems. According to this method, the optimal solution in a multidimensional problem is found by decomposing it into stages, each of which represents a subproblem with respect to one variable. In economic problems, the number of stages is the planning horizon. The choice of a planning horizon is necessary for a rigorous statement of the applied problem in the field of economics and management, but it is often difficult to justify. We see a way out in the use of asymptotically optimal plans for which the values of the optimization criterion differ little from its values for optimal plans for all sufficiently large planning horizons. The main result of the paper is the existence of an asymptotically optimal plan. The proof is carried out in several statements. If the sum of the maximums of the transition functions tends to 0, the existence of an asymptotically optimal plan is obtained in Theorem 1. A special case is models with a discount at a discount coefficient less than 1. The main part of the article is devoted to models with a discount coefficient equal to 1. Theorem 2 on the highway is proved for base set of a finite number of elements. In Theorem 3, a statement is obtained on the approximation of an arbitrary set by a finite one. In the final Theorem 4, the existence of an asymptotically optimal plan is proved in the general case. The term “magistral” is associated with a well-known recommendation to drivers: in order to get from point A to point B, it is advisable to go to the highway (magistral) at the initial section of the road, and then exit the highway and get to point B. The recommendation for choosing the optimal one is similar trajectories using the Pontryagin maximum principle in the model of the optimal distribution of time between obtaining knowledge and developing skills. This fact underlines the methodological proximity of dynamic programming and the Pontryagin maximum principle
-
PROBABILITY-STATISTICAL MODELS OF CORRELATION AND REGRESSION
08.00.13 Mathematical and instrumental methods of Economics
DescriptionThe correlation and determination coefficients are widely used in statistical data analysis. According to measurement theory, Pearson's linear paired correlation coefficient is applicable to variables measured on an interval scale. It cannot be used in the analysis of ordinal data. The nonparametric Spearman and Kendall rank coefficients estimate the relationship of ordinal variables. The critical value when testing the significance of the difference of the correlation coefficient from 0 depends on the sample size. Therefore, using the Chaddock Scale is incorrect. When using a passive experiment, the correlation coefficients are reasonably used for prediction, but not for control. To obtain probabilistic-statistical models intended for control, an active experiment is required. The effect of outliers on the Pearson correlation coefficient is very large. With an increase in the number of analyzed sets of predictors, the maximum of the corresponding correlation coefficients — indicators of approximation quality noticeably increases (the effect of “inflation” of the correlation coefficient). Four main regression analysis models are considered. Models of the least squares method with a determinate independent variable are distinguished. The distribution of deviations is arbitrary, however, to obtain the limit distributions of parameter estimates and regression dependences, we assume that the conditions of the central limit theorem are satisfied. The second type of model is based on a sample of random vectors. The dependence is nonparametric, the distribution of the two-dimensional vector is arbitrary. The estimation of the variance of an independent variable can be discussed only in the model based on a sample of random vectors, as well as the determination coefficient as a quality criterion for the model. Time series smoothing is discussed. Methods of restoring dependencies in spaces of a general nature are considered. It is shown that the limiting distribution of the natural estimate of the dimensionality of the model is geometric, and the construction of an informative subset of features encounters the effect of "inflation coefficient correlation". Various approaches to the regression analysis of interval data are discussed. Analysis of the variety of regression analysis models leads to the conclusion that there is no single “standard model”
-
CHARACTERIZATION OF MODELS WITH DISCOUNTING
08.00.13 Mathematical and instrumental methods of Economics
DescriptionAmong the widely used economic-mathematical models, dynamic programming plays an important role, and among them, models with discounting. The most famous example is the model for calculating the net present value (NPV) as an estimate of the efficiency of the investment project. In the article, it is clarified which features are distinguished by models with discounting among all models of dynamic programming. In models with discounting, the comparison of plans does not change when the time of the beginning of the implementation of plans changes, ie. there is a stability of the results of comparing plans. It is proved that if the results of comparing plans for 1 and 2 steps are stable in the dynamic programming model, then this model is a model with discounting. This theorem shows that the introduction of discounted functions for the estimation of the effect is justified only in stable economic conditions in which the orderliness of managerial decisions does not change from year to year. In other words, if at the beginning of the period under consideration the first solution is better than the second, then at all other times, up to the end of the period under consideration, the first solution is better than the second. Stable economic conditions are rarely found in the modern economy with its constant changes, including those caused by innovations. Therefore, the decision to choose (to implement) an investment project from a set of possible ones can not be based solely on the calculation of discounted project performance indicators, such as net present value and internal rate of return. Such indicators can only play a supporting role. Decide on the choice of an investment project for implementation is necessary on the basis of the whole range of social, technological, environmental, economic, political factors
-
BASIC REQUIREMENTS FOR DATA ANALYSIS METHODS (ON THE EXAMPLE OF CLASSIFICATION TASKS)
08.00.13 Mathematical and instrumental methods of Economics
DescriptionThere is a need to clean up the classification methods. This will increase their role in solving applied problems, in particular, in the diagnosis of materials. For this, first of all, it is necessary to develop requirements that classification methods must satisfy. The initial formulation of such requirements is the main content of this work. Mathematical classification methods are considered as part of the applied statistics methods. The natural requirements to the considered methods of data analysis and the presentation of calculation results arising from the achievements and ideas accumulated by the national probabilistic and statistical scientific school are discussed. Concrete recommendations are given on a number of issues, as well as criticism of individual errors. In particular, data analysis methods must be invariant with respect to the permissible transformations of the scales in which the data are measured, i.e. methods should be adequate in the sense of measurement theory. The basis of a specific statistical method of data analysis is always one or another probabilistic model. It should be clearly described, its premises justified - either from theoretical considerations, or experimentally. Data processing methods intended for use in real-world problems should be investigated for stability with respect to the tolerances of the initial data and model premises. The accuracy of the solutions given by the method used should be indicated. When publishing the results of statistical analysis of real data, it is necessary to indicate their accuracy (confidence intervals). As an estimate of the predictive power of the classification algorithm, it is recommended to use predictive power instead of the proportion of correct forecasts. Mathematical research methods are divided into "exploratory analysis" and "evidence-based statistics." Specific requirements for data processing methods arise in connection with their "docking" during sequential execution. The article discusses limits of applicability of probabilistic-statistical methods. Concrete statements of classification problems and typical errors when applying various methods for solving them are also considered
-
THE ASYMPTOTIC INFORMATION CRITERION OF NOISE QUALITY
DescriptionIntuitively everyone understands that noise is a signal in which there no information is, or which in practice fails to reveal the information. More precisely, it is clear that a certain sequence of elements (the number) the more is the noise, the less information is contained in the values of some elements on the values of others. It is even stranger, that noone has suggested the way, but even the idea of measuring the amount of information in some fragments of signal of other fragments and its use as a criterion for assessing the degree of closeness of the signal to the noise. The authors propose the asymptotic information criterion of the quality of noise, and the method, technology and methodology of its application in practice. As a method of application of the asymptotic information criterion of noise quality, we offer, in practice, the automated systemcognitive analysis (ASC-analysis), and as a technology and software tools of ASC-analysis we offer the universal cognitive analytical system called "Eidos". As a method, we propose a technique of creating applications in the system, as well as their use for solving problems of identification, prediction, decision making and research the subject area by examining its model. We present an illustrative numerical example showing the ideas presented and demonstrating the efficiency of the proposed asymptotic information criterion of the quality of the noise, and the method, technology and methodology of its application in practice