Examine the Reliability of Econometrics Software: An Empirical Comparison of Time Series Modelling

Abstract

Researchers must understand that naively relying on the reliability of statistical software packages may result in suboptimal, biased, or erroneous results, which affects applied economic theory and the conclusions and policy recommendations drawn from it. To create confidence in a result, several software packages should be applied to the same estimation problem. This study examines the results of three software packages (EViews, R, and Stata) in the analysis of time-series econometric data. The time-series data analysis which presents the determinants of macroeconomic growth of Sri Lanka from 1978 to 2020 has been used. The study focuses on testing for stationarity, cointegration, and significant relationships among the variables. The Augmented Dickey-Fuller and Phillips Perron tests were employed in this study to test for stationarity, while the Johansen cointegration test was utilized to test for cointegration. The study employs the vector error correction model to assess the short-run and long-term dynamics of the variables in an attempt to determine the relationship between them. Finally, the Granger Causality test is employed in order to examine the linear causation between the concerned variables. The study revealed that the results produced by three software packages for the same dataset and the same lag order vary significantly. This implies that time series econometrics results are sensitive to the software that is used by the researchers while providing different policy implications even for the same dataset. The present study highlights the necessity of further analysis to investigate the impact of software packages in time series analysis of economic scenarios.

Share and Cite:

Wickramasinghe, W. , Athukorala, P. , Senarathne, S. and Yapa, Y. (2023) Examine the Reliability of Econometrics Software: An Empirical Comparison of Time Series Modelling. Open Journal of Statistics, 13, 25-45. doi: 10.4236/ojs.2023.131003.

1. Introduction

Data analysis is a crucial part of research on which many researchers depend to make research results, conclusions, and recommendations. For a research project to be completed successfully, a rational data analysis procedure is vital. With the growing popularity of data analysis, a significant number of technological tools have developed in the academic context. Statistical software (SS) is one of these tools. The advent of SS has undoubtedly contributed significantly to the advancement of research studies in the 21st century.

Experts have shown that SS is a software application that simplifies the computation and presentation of data. If all data is entered properly, SS permits researchers to avoid frequent numerical errors and generate accurate results in their study. The development of SS enables scholarly researchers to quickly undertake more quantitative analyses. Many academics, experts, scientists, and business managers can also clearly convey accurate future predictions utilizing SS.

Modern statistical software has transformed statistical processes from time-consuming, arduous hand-calculation to working swiftly with large, complex datasets [1] . Incorporating statistical software into statistics education can assist students to enhance their statistical knowledge and enthusiasm toward statistics [2] [3] [4] . There are several proprietary and freeware SS programs available that are useful for various statistical analyses, depending on the demands of the user.

Researchers frequently attempt to uncover flaws in the statistical procedures, instead of treating the software as a possible source of error. Furthermore, many economists think that analyzing the reliability of software packages should be left to software developers, as the economics profession is less experienced in this area. According to McCullough & Vinod [5] , this has made software flaws more frequent than they should be. Economists often place more importance on quickness and usability, but accuracy and reliability may suffer as a result. Moreover, they claim that as a result of these flaws, historical inferences are brought into question, and future work must document and archive statistical software in addition to statistical models.

Software testing research leads to better software. Software accuracy would be neglected if such research did not exist. The significant enhancements of statistical software show that developers are concerned about the reliability of their statistical software package. The fact that software developers, such as Stata and JMP, have posted their product’s reliability on the web is a clear indication that vendors are paying attention. Stata Corporation rectified flaws reported in their benchmarking in response to an earlier conference-paper version of their study, according to Altman & McDonald [6] . While significant advances in computer technology have been achieved, it is still necessary to assess the reliability of statistical software packages on a regular basis to verify whether researchers are obtaining reliable and accurate estimates. According to studies, available software packages are not flawless, and may not be as efficient and accurate as some researchers believe [7] [8] .

McCullough [9] evaluated versions 6.12, 7.5, and 4.0 of SAS, SPSS and S-Plus, particularly focusing on three areas: estimation (linear and nonlinear), random number generation, and statistical distribution. Although there were flaws in all three areas, the packages performed well when calculating univariate statistics and linear regression models. McCullough [10] examined four econometric statistical packages from a study of EViews, LIMDEP, SHAZAM and TSP and the author advised against relying on default nonlinear settings (e.g., algorithmic choice, tolerance level, starting point). Sawitzki [8] tested the numerical reliability of nine packages; BMDP, Data Desk, Excel, Interpretive Software Products (ISP), Generalized Linear Interactive Modelling (GLIM), SAS, Statistical Package for the Social Sciences (SPSS), S-Plus, and Statgraphics, and found that all of them failed basic computational tests like calculating sample variance estimators. McCullough & Heiser [11] investigated the accuracy of various statistical procedures in Excel 2007 and found that it failed accuracy tests for statistical distributions, random number generation, and estimation. L’ecuyer & Simard [12] created a software library for evaluating uniform random number generators and examined a series of frequently used random number generators. They discovered that the default generators in Excel, MATLAB, Mathematica, and R failed many of their tests. McCullough & Wilson [13] investigated the accuracy of statistical methods in Microsoft Excel 97. They found that Excel failed to perform well in linear and nonlinear estimates, random number generation, and statistical distributions (calculation of p-values). The authors advised against using Excel for statistical analyses. McCullough & Wilson [14] investigated the accuracy of statistical processes in Excel 2000 and Excel XP in a follow-up study. The authors noted that issues discovered by Sawitzki [8] in Excel 4.0, such as the instability of the sample variance estimator, were still present in Excel 2000 and Excel XP. McCullough & Wilson [7] indicated that, while there was some progress in some areas of Excel’s computations, errors persisted in its updated versions. McCullough & Wilson [14] demonstrated in another study that issues highlighted in earlier versions of Excel were not resolved in Excel 2003. They noted that the software was still ineffective at handling nonlinear problems and warned users not to use it. McCullough [15] evaluated EViews 3.0, LIMDEP 7.0, RATS (Regression Analysis of Time Series) 4.3, SHAZAM 8.0, and TSP 4.4 for numerical accuracy. He found that, while the use of reliability testing led software developers to enhance these packages, numerous flaws still remained. He demanded that the updated versions of these packages be examined to see if the identified issues have been resolved. Keeling & Pavur [16] compared nine statistical software packages; SAS 9.1, SPSS 12.0, Excel 2003, Minitab 14.0, Stata 8.1, S-Plus 6.2, R 1.9.1, JMP 5.0, and StatCrunch 3.0 in the NIST benchmark tests with regard to univariate summary statistics, analysis of variance, linear regression, and nonlinear regression. They noted that when compared to earlier versions, newer versions of several applications like Excel 2003, SAS, SPSS, and S-Plus had made significant enhancements. They did, however, emphasize that errors remained in some of the packages.

According to the literature reviewed above, users should be aware of the limitations and shortcomings of different software packages. Furthermore, rather than depending on a single package, applying multiple packages and accepting a solution when the results agree across packages may be a more reliable approach. The useful contribution of this study is that it brings attention to the body of literature on variation in outputs by investigating software reliability across three statistical software packages (EViews, R and Stata) for econometric tests.

2. Methodology

Using annual time series data from 1978 to 2020, the study tries to identify the primary causes of growth for the Sri Lankan economy. The study’s data came from the Special Statistical Appendix of the Central Bank of Sri Lanka’s Annual Report 2020, which was accessed online. The real gross domestic product (RGDP), the exchange rate (ER), which signifies the exchange rates from USD to LKR, real public debt (RPD), real government expenditure (RGE), real imports (RI), real exports (RE), and broad money supply M2 (BMS) are among the variables considered in the study. The variables are all expressed in Sri Lankan rupees. Here, the variable RGDP is considered as a proxy to measure the economic growth of Sri Lanka.

R G D P = f ( E R , R P D , R G E , R I , R E , B M S ) (1)

As such, the following Equation (2) is proposed to establish the relationship between economic growth and macroeconomic factors in Sri Lanka. According to the Equation (1), economic growth is affected by the exchange rate, real public debt, real government expenditure, real imports, real exports and broad money supply. Theory suggests that changes in the exchange rate, real public debt, real government expenditure, real imports, real exports and broad money supply will cause a change in real GDP. As a result of the aforementioned theoretical framework, the Sri Lankan economic growth function as it applies to this study may be stated in logarithmic linear form as follows:

ln R G D P t = β 0 + β 1 ln E R t + β 2 ln R P D t + β 3 ln R G E t + β 4 ln R I t + β 5 ln R E t + β 6 ln B M S t + ε t (2)

where lnRGDPt represents the log of real GDP at time t;

lnERt represents the log of exchange rate at time t;

lnRPDt represents the log of real public debt at time t;

lnRGEt represents the log of real government expenditure at time t;

lnRIt represents the log of real imports at time t;

lnREt represents the log of real exports at time t;

lnBMSt represents the log of broad money supply at time t;

t = time

εt is the error term, which is assumed to be normally and independently distributed with a zero mean and constant variance which represents any other explanatory variables that impact economic development but are not represented by this model. In the model β0 is a constant and β1, β2, β3, β4, β5, β6 denotes the sensitivities of each macroeconomic variable to economic growth. The study utilizes the VECM to evaluate the short-run and long-term dynamics of the variables in an attempt to identify the relationship between the variables. All preliminary tests for time series data were performed: test for stationarity using the ADF test and PP test, optimal lag length selection using the AIC, test for cointegration using the Johansen cointegration tests [17] , later our model was fitted using the Johansen normalization technique. And finally, the Granger Causality test was carried out to determine the direction of causality between the variables. Analysis tests were performed using EViews 12.0, R 4.1.2, and Stata/SE 12.0 software.

2.1. Econometric Views (EViews)

EViews is a windows OS statistics tool that is mostly used for time-series related econometrics research. It was created by Quantitative Micro Software (QMS), which is now part of IHS. The most recent version of EViews is 12.0, which was published in November 2020. EViews can be used for general statistical and economic studies such as cross-section and panel data analysis, as well as time series estimation and forecasting. For data storage, EViews mainly depends on a proprietary and undocumented file format. However, it supports a wide range of input and output formats, including databank format, MS-Excel format, SPSS/PSPP, DAP/SAS, Stata, RATS, and TSP. Moreover, EViews can connect to Open Database Connectivity (ODBC) databases.

2.2. STATA

Stata is a robust statistical software that includes intelligent data-management features, a large range of up-to-date statistical methods, and an outstanding system for creating publication quality graphics. The most recent version Stata 17, released on April 20, 2021, is a quick and simple data management program. Stata is available on Windows, Unix, and Mac operating systems. Intercooled Stata (Stata/IC) is the standard version, and it can handle up to 2047 variables. Stata/SE is a special edition that can handle up to 32,766 variables (allows longer string variables and bigger matrices), while Stata/MP is a version for multicore/multiprocessor systems that has the same restrictions but is much faster. Stata conducts the majority of general statistical analyses including regression, logistic regression, survival analysis, analysis of variance, factor analysis, multivariate analysis and time series analysis.

2.3. R

R is a programming language for statistical computing developed by R development core team and the first official stable version was released in 2000. R is a free open-source software, which allows researchers to modify and develop packages as per their requirements. Another advantage of R is that R base environment can be run on multiple IDEs (RStudio, Jupyter notebook, etc.) based on the user preference. The comprehensive R archive network (CRAN) contains more than 10,000 different packages oriented at various different requirements. R also has special packages such as RShiny and ROrcale which enable users to build interactive webapps containing visualizations and interact with databases such as ROracle. The markdown package in R is also popular among users, as it helps to generate reports in the desired formats such as webpages, Word documents, PowerPoint slides, pdfs, etc. The latest version of R as of the date January 22, 2023 is the 4.2.2 version and at present it has grown to be one of the most popular statistical programming environments among researchers and analysts owing to the abovementioned distinguishing features.

3. Results and Discussion

Table 1 displays a descriptive statistic for each variable of interest. The average value of real GDP is Rs. 1,208,842 with the maximum and minimum values being Rs. 2,916,836 and Rs. 309,166 respectively during the study period. The average value of the real public debt is Rs. 1,028,986 with a maximum of Rs. 2,839,453 and a minimum of Rs. 221,713 and the dispersion from the mean value is Rs. 676,202. The average exchange rate (LKR to USD) is Rs. 78.63 while the maximum is Rs. 185.52 and the minimum is Rs. 15.57, with a wider dispersion from the mean of Rs. 49.07. Relative low value of the standard deviation of real export shows a less diversification of the export sector in the country. There is a significant growth of the import as well as money supply in the country during

Table 1. Summary statistics of variables used.

this period. Having described the characteristics of the data and based on their background, we begin the econometric analysis by testing the order of integration using the ADF and PP unit root tests, and then proceed to test for cointegration using the Johansen Cointegration method, followed by testing for causality using the Granger Causality method [18] .

3.1. Unit Root Tests

In this section the time-series properties of the data are examined during the period of 1978-2020, consequently ADF and PP tests are utilized. This testing is necessary to avoid spurious regression, which is a typical issue when estimating a regression line using data whose generated process follows a time trend. These unit-root tests were run in EViews, R and Stata on both levels and first differences of all variables. Furthermore, in order for the tests to be effective, the lag order must be chosen carefully so that the power of the test is not diminished [14] . Accordingly, the lag order was chosen using the minimum AIC values through EViews [19] . The R program takes the default value of

t r u n c { ( l e n g t h ( x ) 1 ) } 1 3 as the lag order for ADF test, where x denotes the

variable of interest. The Stata program allows the user to specify the number of lags to be used in the unit root tests. Both AIC and SBIC are often used to determine the optimal lag length. The aim is to choose the number of parameters that minimize the value of the information criterion. The SBIC has a propensity to underestimate the lag order, whereas adding more lags raises the penalty for the loss of degrees of freedom. Hence, AIC is used as the leading indicator to ensure that there is no residual autocorrelation in the VAR model. To study the discrepancies in the results, the three programs were run in the same lag order. Table 2 shows the results of the ADF test performed by the three programs.

Table 2 indicates that for the three programs, all variables have a unit root at their level for the ADF test, and the p values for all series are not significant at the 1% and 5% significance levels. However, when the unit root test is run at the first difference, the ADF test results showed that all variables are stationary at the first difference, I (1), and the p-values are significant at 1% and 5% in all three programs. Based on these estimated results, we failed to reject the null hypothesis of unit roots at all level, but we reject the presence of unit root at all first differences of the variables. As a result, the variables are integrated of order 1, I (1). Table 3 shows the results of the PP test performed by the three programs.

To test each series for unit roots, Phillips–Perron test utilized the Bartlett Kernel with automated Newey–West bandwidth selection. Similar to the ADF test results, the variables in the levels are nonstationary, but their first differences are stationary, I (1). Moreover, the test statistics for the ADF and PP tests in EViews and Stata are nearly identical, although the test statistics provided by the R application are distinct. The default lag value for the PP test in Stata is computed using i n t ( 4 ( N / 100 ) 2 / 9 ) . To examine the variations in the outcomes, the same lag order is used to run the PP test in all three applications, as determined by the Bartlett Kernel with automatic Newey-West bandwidth selection through EViews. The results of the preceding unit-root tests reveal that all the seven variables are nonstationary in levels, but become stationary after taking their first difference, indicating that all the variables are integrated of order one, I (1) series. It is vital to emphasize that differencing a variable to attain stationarity might result in long run information loss. Thus, we perform a cointegration test to determine whether there is a long run relationship between the dependent and explanatory variables in various models.

Table 2. Augmented dickey-fuller unit root test for stationarity with intercept only.

a. MacKinnon (1996) one-sided p-values for rejection of hypothesis of unit root are denoted inside parentheses. *Denotes significance at the 10% level. **denotes significance at the 5% level. ***denotes significance at the 1% level.

Table 3. Phillips perron unit root test for stationarity with intercept only.

a. MacKinnon (1996) critical values for rejection of hypothesis of unit root are denoted in parentheses. *Denotes significance at the 10% level. **denotes significance at the 5% level. ***denotes significance at the 1% level. The spectral estimation is done using the Bartlett kernel with Newey–West automatic bandwidth. Bandwidth is reported inside square brackets.

3.2. Cointegration Tests

The Johansen Maximum cointegration test [17] is applied based on the above unit-root tests. Before conducting cointegration tests, the relevant order of lags (p) of the VAR model must be specified. Based on the minimum AIC values, EViews suggests lag order 2 for the VAR model. The same lag order is utilized in R and Stata software to conduct cointegration tests. However, in the R application, the information criteria specified lag order 5 under the VAR select function. EViews and Stata software failed to produce acceptable results for the Johansen cointegration test at lag 5. Furthermore, despite lag selection using the R application, the first two test statistics of r = 0 and r ≤ 1 were generated as NaN. Consequently, lag 2 was used in all three applications for the test.

The results of the maximum eigenvalue and trace tests are shown in Table 4, starting with the null hypothesis that there is no cointegration (r = 0) among the seven variables. The maximum eigenvalue and the trace statistic in both EViews and R suggest r = 2. As a result, it is inferred that the variables have two cointegrating relationships. In Stata, the maximum eigenvalue and trace statistic indicate r = 1, implying that there is only one cointegrating relationship among the variables. This result clearly shows that different statistical programs generate the different results for econometric modelling of macroeconomic variables. Table 5 shows estimates of long-run cointegrating vectors.

Empirical results of the long run relationship taken from EViews, R and Stata with their significance level are summarized in Table 5. Interestingly, R application does not provide the information that are needed to decide significance of the variables. According to Engle & Granger [20] , Marquez [21] testing the level of significance of the standardized cointegration vector is not required.

Table 4. Johansen cointegration tests.

a. *denotes rejection of hypothesis at the 0.05 level.

Table 5. Normalized cointegration vectors.

a. *Denotes significance at the 10% level. **denotes significance at the 5% level; ***denotes significance at the 1% level.

The EViews and R software generated two normalized cointegration vectors, whereas Stata generated one normalized cointegration vector. In each of the three applications, there is a distinct variation in the coefficients of the long run relationship. The long run models can be re-parameterized as follows.

Table 6 demonstrates a significant difference in the long run models generated by the three programs. Despite the fact that the Johansen cointegration results of R program revealed the existence of two cointegration equations, the program produced only one normalized cointegration vector for the variable RGDP. The signs of the coefficients of the variables in the long run models created by R and Stata are identical, which is a distinguishing feature of the long run models developed by normalizing for real GDP variable; however, the signs of the variables lnRI and lnRE are opposite in EViews when compared with the other two programs. Moreover, coefficient of the lnRE variable produced by EViews is 1.462 whereas the coefficient produced by R and Stata are −0.842 and −0.928 respectively. The coefficients produced for lnRPD are −3.12, −2.257 and −1.892 by EViews, R and Stata respectively, suggesting a negative impact from real public debt towards real GDP in the long run. The coefficients produced by EViews, R, and Stata for lnER are 0.072, 0.999, and 0.891 respectively, indicating that the exchange rate has a positive impact on real GDP in the long run. The existence of cointegration necessitates the estimation of an ECM, which represents the short-run dynamics. Therefore, in order to fully investigate the impact of macroeconomic variables on economic growth, short-run estimations were performed in this paper that is discussed in the next section.

Table 6. Long run models of the three software.

3.3. Vector Error Correction Model (VECM)

The VECM is used to investigate the short run dynamics of the variables. The error correction model allows previous disequilibrium to be introduced as independent variables in the dynamic behavior of existing variables. VECM relates changes in RGDP growth to changes in the other lagged variables and the disturbance term of lagged periods. Table 7 shows the results of vector error correction model’s short run dynamic relationship and set of short run coefficients for the three software.

The EViews results show that the coefficient of the ECT is negative and is not significant at 5 percent. Additionally, the immediate impact of the explanatory variables demonstrates that the last two years of real public debt and real exports had negative impacts on the growth of real GDP and the last year of real government expenditure and broad money supply M2 had negative and positive impacts on real GDP respectively. These impacts were statistically significant. Therefore, an increase in the past two years of real public debt will cause growth in real GDP to decline by 0.24 percent while an increase in the past year of broad money supply will cause growth in real GDP to increase by 0.35 percent. Moreover, an increase in the past year of real government expenditure will cause the growth of real GDP to fall by 0.27 percent. Alternatively, the past two-year records of growth in real GDP had a significant positive impact on current growth in real GDP. Therefore, in the short run, real public debt, real government expenditure, real exports and broad money supply are determinants of growth of real GDP.

In contrast, the R results show that the coefficient of the ECT is negative and significant at 1 percent. This implies that each year, a 15.8 percentage point modification is made to the long run periods. Table 7 also indicates that the past year of broad money supply had a negative significant impact on growth of real GDP. Therefore, an increase in the broad money supply in the past year will cause the growth of real GDP to decrease by 0.3 percent. The past records of exchange rate, real government expenditure and real exports had a negative impact on growth in real GDP. Moreover, the past records of real imports had a positive impact on the growth of real GDP. However, the impacts of these variables are not statistically significant.

The coefficient of the EC term is negative and significant at 1 percent, according to the Stata results. This means that the long run periods are modified by 24.8 percentage points per year. The past year of real public debt, real exports and broad money supply had a positive significant impact on the growth of real GDP. Furthermore, records of real government expenditure and real imports in the past year had a significant negative impact on the growth of real GDP. Therefore, an increase in the real public debt in the past year will cause the growth in real GDP to increase by 0.34 percent. An increase in the past year of real government expenditure will cause the growth in real GDP to decline by 0.25 percent. Whereas a rise in real imports in the previous year caused a 0.11 percent drop in real GDP growth. Moreover, an increase in real exports over the last year will result in a 0.18 growth in real GDP. Hence, in the short run model produced by Stata, real public debt, real government expenditure, real imports, real exports and broad money supply are determinants of growth of real GDP.

Although the co-integration test helps detect the long-term equilibrium relationship between variables, it cannot identify the causal relationship between variables. Therefore, Granger causality test is employed to determine the causal relationship between the variables.

Table 7. VECM Estimation for lnRGDP.

a. *Denotes significance at the 10% level. **denotes significance at the 5% level; ***denotes significance at the 1% level.

3.4. Granger Causality Test

Granger [22] causality test has been performed in order to examine the linear causation between the concerned variables. Granger causality is useful in determining the direction of the relationships. The presence of a cointegration vector, according to Granger, indicates that Granger causality must occur in at least one direction. Tables 8-10 exhibit the results of the Granger causality test conducted by EViews, Stata, and R, respectively, using the optimal lag length of 1 based on AIC.

From Table 8, the result shows that there is unilateral directional causality between real GDP and real imports, exchange rate and real government expenditure, real imports and exchange rate, real government expenditure and real imports, real exports and real government expenditure. Also, there is bidirectional causality between broad money supply and real imports.

Table 9 shows that there is unilateral directional causality between real public debt and real GDP, real government expenditure and real GDP, real imports and real GDP, real exports and real GDP. Table 10 shows unilateral directional causality for real public debt and real GDP, real GDP and real imports, real GDP and real exports, exchange rate and real public debt, exchange rate and real imports, exchange rate and real exports, and bidirectional causality between real GDP and real government expenditure. Accordingly, it can be concluded that there is a significant variation of the results produced by different statistical packages in time series modelling in the field of macroeconomics. The results generated by econometric analyses should be interpreted and explained in the context of economic theory and principles [4] . This will lead bias results and any policies drawn from the results taken from using different computer packages may mislead the true behavior of the macroeconomic variables.

Table 8. Granger causality test results EViews.

Table 9. Granger causality test results Stata.

Table 10. Granger causality test results R.

4. Conclusions and Future Work

The primary goal of this study is to examine the discrepancies in the results of econometric analysis tests performed by EViews, R and Stata on the same dataset. The procedures used in all three programs are similar and the same data set is used. To test the stationarity and the order of integration of all series, the ADF test and the PP test were employed and Akaike Information Criterion was used to find the optimal lag structure for each variable. The results of these tests indicated that all the variables used in the study are integrated into order 1. The test statistics of the ADF test and PP test produced by EViews and Stata programs are nearly identical for the same lag order and the test statistic of the tests produced by R are distinct when compared with EViews and Stata. Later an econometric analysis was conducted based on the Johansen cointegration test and a vector error correction model to analyze the interrelationships between macroeconomic variables and economic growth. The Johansen cointegration tests carried out in EViews and R programs inferred that there exists a long run equilibrium relationship between the variables and the presence of two cointegrating equations at a 0.05 significance level. The Johansen test procedure run in the Stata program confirmed that there is at least one cointegrating equation at 0.05 significance for the economic growth and macroeconomic variables. The coefficients of the long run model produced by the Johansen test in the three programs show a significant variation when compared with each other. The signs of the variables in the long-run model produced by R and Stata are identical when compared with the EViews program. In the long run, economic growth is positively impacted by exchange rate, real government expenditure and broad money supply, while real public debt negatively impacts the economic growth. In the R and Stata programs, real imports and real exports had a positive and negative impact on economic growth, respectively. Whereas in EViews, the impact of real imports and real exports were negative and positive on economic growth, respectively.

To assess the short run dynamics of the variables vector error correction mechanism was carried out and the results of the VEC in EViews show that short run relationships exist between real GDP, real public debt, real government expenditure, real exports and broad money supply. The R results of VEC show that there exists a short-run relationship between real GDP and broad money supply. The VEC run in the Stata program indicated that there exist significant short run relationships between real GDP, real public debt, real government expenditure, real imports, real exports and broad money supply. The Granger causality test findings also revealed differences in the three software. EViews and R showed that there is a unilateral directional causality between real GDP and real imports. The results of EViews and Stata yielded a similar conclusion for unilateral directional causation between broad money supply and real GDP. For unilateral directional causations between broad money supply with real imports, real exports and real government expenditure, the Stata and R findings yielded the same conclusion.

Taken altogether, these findings suggest that the statistical econometric analyses performed by different applications yield diverse results. This paper provides new insights related to the results of tests performed in different applications. The main explanation for the disparities in results might be changes in the theoretical formulations of the tests listed. More investigation should be conducted to figure out the actual cause of these disparities in results across different applications.

Acknowledgements

This research received no specific grant from any funding agency in the public, commercial, or not-for-profit sectors.

Appendix

Glossary of Abbreviations

ADF Augmented Dickey-Fuller

AIC Akaike Information Criteria

CRAN Comprehensive R Archive Network

GDP Gross Domestic Product

IHS Information Handling Services

ISP Interpretive Software Products

JMP John’s Macintosh Project

LKR Sri Lankan Rupee

NIST National Institute of Standards and Technology

ODBC Open Database Connectivity

OS Operating System

PP Phillips Perron

QMS Quantitative Micro Software

RATS Regression Analysis of Time Series

SBIC Schwarz Bayesian Information Criteria

SPSS Statistical Package for the Social Sciences

SS Statistical Software

TSP Time Series Processor

USD United States Dollar

VAR Vector Autoregressive

VECM Vector Error Correction Model

Conflicts of Interest

The authors declare no conflicts of interest.

References

[1] Chen, E.E. and Wojcik, S.P. (2016) A Practical Guide to Big Data Research in Psychology. Psychological Methods, 21, 458-474.
https://doi.org/10.1037/met0000111
[2] Ciftci, S.K., Karadag, E. and Akdal, P. (2014) Instruction of Statistics via Computer-Based Tools: Effects on Statistics’ Anxiety, Attitude, and Achievement. Journal of Educational Computing Research, 50, 119-133.
https://doi.org/10.2190/EC.50.1.f
[3] McCulloch, R S. (2017) Learning Outcomes in a Laboratory Environment vs. Classroom for Statistics Instruction: An Alternative Approach Using Statistical Software. International Journal of Higher Education, 6, 131-142.
https://doi.org/10.5430/ijhe.v6n5p131
[4] Sosa, G.W., Berger, D.E., Saw, A.T. and Mary, J.C. (2011) Effectiveness of Computer-Assisted Instruction in Statistics: A Meta-Analysis. Review of Educational Research, 81, 97-128.
https://doi.org/10.3102/0034654310378174
[5] McCullough, B.D. and Vinod, H.D. (1999) The Numerical Reliability of Econometric Software. Journal of Economic Literature, 37, 633-665.
https://doi.org/10.1257/jel.37.2.633
[6] Altman, M. and McDonald, M.P. (2001) Choosing Reliable Statistical Software. PS: Political Science & Politics, 34, 681-687.
https://doi.org/10.1017/S1049096501001093
[7] McCullough, B.D. and Wilson, B. (2005) On the Accuracy of Statistical Procedures in Microsoft Excel 2003. Computational Statistics & Data Analysis, 49, 1244-1252.
https://doi.org/10.1016/j.csda.2004.06.016
[8] Sawitzki, G. (1994) Report on the Numerical Reliability of Data Analysis Systems. Computational Statistics & Data Analysis, 18, 289-301.
https://doi.org/10.1016/0167-9473(94)90177-5
[9] McCullough, B.D. (1999) Assessing the Reliability of Statistical Software: Part II. The American Statistician, 53, 149-159.
https://doi.org/10.1080/00031305.1999.10474450
[10] McCullough, B.D. (1999) Econometric Software Reliability: EViews, LIMDEP, SHAZAM and TSP. Journal of Applied Econometrics, 14, 191-202.
https://doi.org/10.1002/(SICI)1099-1255(199903/04)14:2<191::AID-JAE524>3.0.CO;2-K
[11] McCullough, B.D. and Heiser, D.A. (2008) On the Accuracy of Statistical Procedures in Microsoft Excel 2007. Computational Statistics & Data Analysis, 52, 4570-4578.
https://doi.org/10.1016/j.csda.2008.03.004
[12] L’Ecuyer, P. and Simard, R. (2007) TestU01: AC Library for Empirical Testing of Random Number Generators. ACM Transactions on Mathematical Software (TOMS), 33, Article No. 22.
https://doi.org/10.1145/1268776.1268777
[13] McCullough, B.D. and Wilson, B. (2002) On the accuracy of statistical procedures in Microsoft Excel 2000 and Excel XP. Computational Statistics & Data Analysis, 40, 713-721.
https://doi.org/10.1016/S0167-9473(02)00095-6
[14] Ng, S. and Perron, P. (2001) LAG Length Selection and the Construction of Unit Root Tests with Good Size and Power. Econometrica, 69, 1519-1554.
https://doi.org/10.1111/1468-0262.00256
[15] McCullough, B.D. (2004) Wilkinson’s Tests and Econometric Software. Journal of Economic and Social Measurement, 29, 261-270.
https://doi.org/10.3233/JEM-2004-0199
[16] Keeling, K.B. and Pavur, R.J. (2007) A Comparative Study of the Reliability of Nine Statistical Software Packages. Computational Statistics & Data Analysis, 51, 3811-3831.
https://doi.org/10.1016/j.csda.2006.02.013
[17] Johansen, S. and Juselius, K. (1990) Maximum Likelihood Estimation and Inference on Cointegration—With Applications to the Demand for Money. Oxford Bulletin of Economics and Statistics, 52, 169-210.
https://doi.org/10.1111/j.1468-0084.1990.mp52002003.x
[18] Dickey, D.A., Jansen, D.W. and Thornton, D.L. (1994) A Primer on Cointegration with an Application to Money and Income. In: Rao, B.B., Ed., Cointegration, Palgrave Macmillan, London, 9-45.
https://doi.org/10.1007/978-1-349-23529-2_2
[19] Shibata, R. (1976) Selection of the Order of an Autoregressive Model by Akaike’s Information Criterion. Biometrika, 63, 117-126.
https://doi.org/10.1093/biomet/63.1.117
[20] Engle, R.F. and Granger, C.W.J. (1987) Co-Integration and Error Correction: Representation, Estimation, and Testing. Econometrica, 55, 251-276.
https://doi.org/10.2307/1913236
[21] Marquez, J. (1995) Time Series Analysis: James D. Hamilton, 1994, (Princeton University Press, Princeton, NJ), 799 pp., US $55.00, ISBN 0-691-04289-6. International Journal of Forecasting, 11, 494-495.
https://doi.org/10.1016/0169-2070(95)90035-7
[22] Granger, C.W.J. (1969) Investigating Causal Relations by Econometric Models and Cross-Spectral Methods. Econometrica: Journal of the Econometric Society, 37, 424-438.
https://doi.org/10.2307/1912791

Copyright © 2024 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.