Gretl Guide
Gretl Guide
Gretl Guide
Allin Cottrell
Department of Economics
Wake Forest University
July, 2019
Permission is granted to copy, distribute and/or modify this document under the terms of the
GNU Free Documentation License, Version 1.1 or any later version published by the Free Software
Foundation (see http://www.gnu.org/licenses/fdl.html).
Contents
1 Introduction 1
1.1 Features at a glance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Acknowledgements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.3 Installing the programs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
2 Getting started 4
2.1 Let’s run a regression . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
2.2 Estimation output . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
2.3 The main window menus . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
2.4 Keyboard shortcuts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
2.5 The gretl toolbar . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
3 Modes of working 12
3.1 Command scripts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
3.2 Saving script objects . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
3.3 The gretl console . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
3.4 The Session concept . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
4 Data files 18
4.1 Data file formats . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
4.2 Databases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
4.3 Creating a dataset from scratch . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
4.4 Structuring a dataset . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
4.5 Panel data specifics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
4.6 Missing data values . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
4.7 Maximum size of data sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
4.8 Data file collections . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
4.9 Assembling data from multiple sources . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
5 Sub-sampling a dataset 31
5.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
5.2 Setting the sample . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
5.3 Restricting the sample . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
i
Contents ii
8 Realtime data 65
8.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
8.2 Atomic format for realtime data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
8.3 More on time-related options . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
8.4 Getting a certain data vintage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
8.5 Getting the n-th release for each observation period . . . . . . . . . . . . . . . . . . . . . 68
8.6 Getting the values at a fixed lag after the observation period . . . . . . . . . . . . . . . . 69
8.7 Getting the revision history for an observation . . . . . . . . . . . . . . . . . . . . . . . . 70
10.2 Series . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
10.3 Scalars . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79
10.4 Matrices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79
10.5 Lists . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79
10.6 Strings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79
10.7 Bundles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80
10.8 Arrays . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84
10.9 The life cycle of gretl objects . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87
11 Discrete variables 90
11.1 Declaring variables as discrete . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90
11.2 Commands for discrete variables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91
12 Loop constructs 95
12.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
12.2 Loop control variants . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
12.3 Progressive mode . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
12.4 Loop examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
32 Forecasting 291
32.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 291
Contents vii
IV Appendices 391
Bibliography 408
Chapter 1
Introduction
User-friendly Gretl offers an intuitive user interface; it is very easy to get up and running with
econometric analysis. Thanks to its association with the econometrics textbooks by Ramu
Ramanathan, Jeffrey Wooldridge, and James Stock and Mark Watson, the package offers many
practice data files and command scripts. These are well annotated and accessible. Two other
useful resources for gretl users are the available documentation and the gretl-users mailing
list.
Flexible You can choose your preferred point on the spectrum from interactive point-and-click to
complex scripting, and can easily combine these approaches.
Cross-platform Gretl’s “home” platform is Linux but it is also available for MS Windows and Mac
OS X, and should work on any unix-like system that has the appropriate basic libraries (see
Appendix C).
Open source The full source code for gretl is available to anyone who wants to critique it, patch it,
or extend it. See Appendix C.
Sophisticated Gretl offers a full range of least-squares based estimators, either for single equations
and for systems, including vector autoregressions and vector error correction models. Sev-
eral specific maximum likelihood estimators (e.g. probit, ARIMA, GARCH) are also provided
natively; more advanced estimation methods can be implemented by the user via generic
maximum likelihood or nonlinear GMM.
Extensible Users can enhance gretl by writing their own functions and procedures in gretl’s script-
ing language, which includes a wide range of matrix functions.
Accurate Gretl has been thoroughly tested on several benchmarks, among which the NIST refer-
ence datasets. See Appendix D.
Internet ready Gretl can fetch materials such databases, collections of textbook datafiles and add-
on packages over the internet.
International Gretl will produce its output in English, French, Italian, Spanish, Polish, Portuguese,
German, Basque, Turkish, Russian, Albanian or Greek depending on your computer’s native
language setting.
1.2 Acknowledgements
The gretl code base originally derived from the program ESL (“Econometrics Software Library”),
written by Professor Ramu Ramanathan of the University of California, San Diego. We are much in
debt to Professor Ramanathan for making this code available under the GNU General Public Licence
and for helping to steer gretl’s early development.
1
Chapter 1. Introduction 2
We are also grateful to the authors of several econometrics textbooks for permission to package for
gretl various datasets associated with their texts. This list currently includes William Greene, au-
thor of Econometric Analysis; Jeffrey Wooldridge (Introductory Econometrics: A Modern Approach);
James Stock and Mark Watson (Introduction to Econometrics); Damodar Gujarati (Basic Economet-
rics); Russell Davidson and James MacKinnon (Econometric Theory and Methods); and Marno Ver-
beek (A Guide to Modern Econometrics).
GARCH estimation in gretl is based on code deposited in the archive of the Journal of Applied
Econometrics by Professors Fiorentini, Calzolari and Panattoni, and the code to generate p-values
for Dickey–Fuller tests is due to James MacKinnon. In each case we are grateful to the authors for
permission to use their work.
With regard to the internationalization of gretl, thanks go to Ignacio Díaz-Emparanza (Spanish),
Michel Robitaille and Florent Bresson (French), Cristian Rigamonti (Italian), Tadeusz Kufel and Pawel
Kufel (Polish), Markus Hahn and Sven Schreiber (German), Hélio Guilherme and Henrique Andrade
(Portuguese), Susan Orbe (Basque), Talha Yalta (Turkish) and Alexander Gedranovich (Russian).
Gretl has benefitted greatly from the work of numerous developers of free, open-source software:
for specifics please see Appendix C. Our thanks are due to Richard Stallman of the Free Software
Foundation, for his support of free software in general and for agreeing to “adopt” gretl as a GNU
program in particular.
Many users of gretl have submitted useful suggestions and bug reports. In this connection par-
ticular thanks are due to Ignacio Díaz-Emparanza, Tadeusz Kufel, Pawel Kufel, Alan Isaac, Cri
Rigamonti, Sven Schreiber, Talha Yalta, Andreas Rosenblad, and Dirk Eddelbuettel, who maintains
the gretl package for Debian GNU/Linux.
MS Windows
The MS Windows version comes as a self-extracting executable. Installation is just a matter of
downloading gretl_install.exe and running this program. You will be prompted for a location
to install the package.
Mac OS X
The Mac version comes as a gzipped disk image. Installation is a matter of downloading the image
file, opening it in the Finder, and dragging Gretl.app to the Applications folder. However, when
installing for the first time two prerequisite packages must be put in place first; details are given
on the gretl website.
1 In this manual we use “Linux” as shorthand to refer to the GNU/Linux operating system. What is said herein about
Linux mostly applies to other unix-type systems too, though some local modifications may be needed.
Part I
3
Chapter 2
Getting started
If you select a row in this window and click on “Info” this opens a window showing information on
the data set in question (for example, on the sources and definitions of the variables). If you find
a file that is of interest, you may open it by clicking on “Open”, or just double-clicking on the file
name. For the moment let’s open data3-6.
+ In gretl windows containing lists, double-clicking on a line launches a default action for the associated list
entry: e.g. displaying the values of a data series, opening a file.
1 For convenience we refer to the graphical client program simply as gretl in this manual. Note, however, that the
specific name of the program differs according to the computer platform. On Linux it is called gretl_x11 while on
MS Windows it is gretl.exe. On Linux systems a wrapper script named gretl is also installed — see also the Gretl
Command Reference.
4
Chapter 2. Getting started 5
This file contains data pertaining to a classic econometric “chestnut”, the consumption function.
The data window should now display the name of the current data file, the overall data range and
sample range, and the names of the variables along with brief descriptive tags — see Figure 2.2.
OK, what can we do now? Hopefully the various menu options should be fairly self explanatory. For
now we’ll dip into the Model menu; a brief tour of all the main window menus is given in Section 2.3
below.
Gretl’s Model menu offers numerous various econometric estimation routines. The simplest and
most standard is Ordinary Least Squares (OLS). Selecting OLS pops up a dialog box calling for a
model specification —see Figure 2.3.
To select the dependent variable, highlight the variable you want in the list on the left and click
the arrow that points to the Dependent variable slot. If you check the “Set as default” box this
variable will be pre-selected as dependent when you next open the model dialog box. Shortcut:
double-clicking on a variable on the left selects it as dependent and also sets it as the default. To
select independent variables, highlight them on the left and click the green arrow (or right-click the
Chapter 2. Getting started 6
highlighted variable); to remove variables from the selected list, use the rad arrow. To select several
variable in the list box, drag the mouse over them; to select several non-contiguous variables, hold
down the Ctrl key and click on the variables you want. To run a regression with consumption as
the dependent variable and income as independent, click Ct into the Dependent slot and add Yt to
the Independent variables list.
The output window contains menus that allow you to inspect or graph the residuals and fitted
values, and to run various diagnostic tests on the model.
For most models there is also an option to print the regression output in LATEX format. See Chap-
ter 38 for details.
To import gretl output into a word processor, you may copy and paste from an output window,
using its Edit menu (or Copy button, in some contexts) to the target program. Many (not all) gretl
windows offer the option of copying in RTF (Microsoft’s “Rich Text Format”) or as LATEX. If you are
pasting into a word processor, RTF may be a good option because the tabular formatting of the
output is preserved.2 Alternatively, you can save the output to a (plain text) file then import the
file into the target program. When you finish a gretl session you are given the option of saving all
the output from the session to a single file.
Note that on the gnome desktop and under MS Windows, the File menu includes a command to
send the output directly to a printer.
+ When pasting or importing plain text gretl output into a word processor, select a monospaced or typewriter-
style font (e.g. Courier) to preserve the output’s tabular formatting. Select a small font (10-point Courier
should do) to prevent the output lines from being broken in the wrong place.
2 Note that when you copy as RTF under MS Windows, Windows will only allow you to paste the material into ap-
plications that “understand” RTF. Thus you will be able to paste into MS Word, but not into notepad. Note also that
there appears to be a bug in some versions of Windows, whereby the paste will not work properly unless the “target”
application (e.g. MS Word) is already running prior to copying the material in question.
Chapter 2. Getting started 7
• File menu
– Open data: Open a native gretl data file or import from other formats. See Chapter 4.
– Append data: Add data to the current working data set, from a gretl data file, a comma-
separated values file or a spreadsheet file.
– Save data: Save the currently open native gretl data file.
– Save data as: Write out the current data set in native format, with the option of using
gzip data compression. See Chapter 4.
– Export data: Write out the current data set in Comma Separated Values (CSV) format, or
the formats of GNU R or GNU Octave. See Chapter 4 and also Appendix E.
– Send to: Send the current data set as an e-mail attachment.
– New data set: Allows you to create a blank data set, ready for typing in values or for
importing series from a database. See below for more on databases.
– Clear data set: Clear the current data set out of memory. Generally you don’t have to do
this (since opening a new data file automatically clears the old one) but sometimes it’s
useful.
– Script files: A “script” is a file containing a sequence of gretl commands. This item
contains entries that let you open a script you have created previously (“User file”), open
a sample script, or open an editor window in which you can create a new script.
– Session files: A “session” file contains a snapshot of a previous gretl session, including
the data set used and any models or graphs that you saved. Under this item you can
open a saved session or save the current session.
– Databases: Allows you to browse various large databases, either on your own computer
or, if you are connected to the internet, on the gretl database server. See Section 4.2 for
details.
– Exit: Quit the program. You’ll be prompted to save any unsaved work.
• Tools menu
– Statistical tables: Look up critical values for commonly used distributions (normal or
Gaussian, t, chi-square, F and Durbin–Watson).
– P-value finder: Look up p-values from the Gaussian, t, chi-square, F, gamma, binomial or
Poisson distributions. See also the pvalue command in the Gretl Command Reference.
– Distribution graphs: Produce graphs of various probability distributions. In the resulting
graph window, the pop-up menu includes an item “Add another curve”, which enables
you to superimpose a further plot (for example, you can draw the t distribution with
various different degrees of freedom).
– Test statistic calculator: Calculate test statistics and p-values for a range of common hy-
pothesis tests (population mean, variance and proportion; difference of means, variances
and proportions).
– Nonparametric tests: Calculate test statistics for various nonparametric tests (Sign test,
Wilcoxon rank sum test, Wilcoxon signed rank test, Runs test).
– Seed for random numbers: Set the seed for the random number generator (by default
this is set based on the system time when the program is started).
Chapter 2. Getting started 8
– Command log: Open a window containing a record of the commands executed so far.
– Gretl console: Open a “console” window into which you can type commands as you would
using the command-line program, gretlcli (as opposed to using point-and-click).
– Start Gnu R: Start R (if it is installed on your system), and load a copy of the data set
currently open in gretl. See Appendix E.
– Sort variables: Rearrange the listing of variables in the main window, either by ID number
or alphabetically by name.
– Function packages: Handles “function packages” (see Section 13.5), which allow you to
access functions written by other users and share the ones written by you.
– NIST test suite: Check the numerical accuracy of gretl against the reference results for
linear regression made available by the (US) National Institute of Standards and Technol-
ogy.
– Preferences: Set the paths to various files gretl needs to access. Choose the font in which
gretl displays text output. Activate or suppress gretl’s messaging about the availability
of program updates, and so on. See the Gretl Command Reference for further details.
• Data menu
– Select all: Several menu items act upon those variables that are currently selected in the
main window. This item lets you select all the variables.
– Display values: Pops up a window with a simple (not editable) printout of the values of
the selected variable or variables.
– Edit values: Opens a spreadsheet window where you can edit the values of the selected
variables.
– Add observations: Gives a dialog box in which you can choose a number of observations
to add at the end of the current dataset; for use with forecasting.
– Remove extra observations: Active only if extra observations have been added automati-
cally in the process of forecasting; deletes these extra observations.
– Read info, Edit info: “Read info” just displays the summary information for the current
data file; “Edit info” allows you to make changes to it (if you have permission to do so).
– Print description: Opens a window containing a full account of the current dataset, in-
cluding the summary information and any specific information on each of the variables.
– Add case markers: Prompts for the name of a text file containing “case markers” (short
strings identifying the individual observations) and adds this information to the data set.
See Chapter 4.
– Remove case markers: Active only if the dataset has case markers identifying the obser-
vations; removes these case markers.
– Dataset structure: invokes a series of dialog boxes which allow you to change the struc-
tural interpretation of the current dataset. For example, if data were read in as a cross
section you can get the program to interpret them as time series or as a panel. See also
section 4.4.
– Compact data: For time-series data of higher than annual frequency, gives you the option
of compacting the data to a lower frequency, using one of four compaction methods
(average, sum, start of period or end of period).
– Expand data: For time-series data, gives you the option of expanding the data to a higher
frequency.
– Transpose data: Turn each observation into a variable and vice versa (or in other words,
each row of the data matrix becomes a column in the modified data matrix); can be useful
with imported data that have been read in “sideways”.
• View menu
Chapter 2. Getting started 9
– Icon view: Opens a window showing the content of the current session as a set of icons;
see section 3.4.
– Graph specified vars: Gives a choice between a time series plot, a regular X–Y scatter
plot, an X–Y plot using impulses (vertical bars), an X–Y plot “with factor separation” (i.e.
with the points colored differently depending to the value of a given dummy variable),
boxplots, and a 3-D graph. Serves up a dialog box where you specify the variables to
graph. See Chapter 6 for details.
– Multiple graphs: Allows you to compose a set of up to six small graphs, either pairwise
scatter-plots or time-series graphs. These are displayed together in a single window.
– Summary statistics: Shows a full set of descriptive statistics for the variables selected in
the main window.
– Correlation matrix: Shows the pairwise correlation coefficients for the selected variables.
– Cross Tabulation: Shows a cross-tabulation of the selected variables. This works only if
at least two variables in the data set have been marked as discrete (see Chapter 11).
– Principal components: Produces a Principal Components Analysis for the selected vari-
ables.
– Mahalanobis distances: Computes the Mahalanobis distance of each observation from
the centroid of the selected set of variables.
– Cross-correlogram: Computes and graphs the cross-correlogram for two selected vari-
ables.
• Add menu Offers various standard transformations of variables (logs, lags, squares, etc.) that
you may wish to add to the data set. Also gives the option of adding random variables, and
(for time-series data) adding seasonal dummy variables (e.g. quarterly dummy variables for
quarterly data).
• Sample menu
– Set range: Select a different starting and/or ending point for the current sample, within
the range of data available.
– Restore full range: self-explanatory.
– Define, based on dummy: Given a dummy (indicator) variable with values 0 or 1, this
drops from the current sample all observations for which the dummy variable has value
0.
– Restrict, based on criterion: Similar to the item above, except that you don’t need a pre-
defined variable: you supply a Boolean expression (e.g. sqft > 1400) and the sample is
restricted to observations satisfying that condition. See the entry for genr in the Gretl
Command Reference for details on the Boolean operators that can be used.
– Random sub-sample: Draw a random sample from the full dataset.
– Drop all obs with missing values: Drop from the current sample all observations for
which at least one variable has a missing value (see Section 4.6).
– Count missing values: Give a report on observations where data values are missing. May
be useful in examining a panel data set, where it’s quite common to encounter missing
values.
– Set missing value code: Set a numerical value that will be interpreted as “missing” or “not
available”. This is intended for use with imported data, when gretl has not recognized
the missing-value code used.
• Variable menu Most items under here operate on a single variable at a time. The “active”
variable is set by highlighting it (clicking on its row) in the main data window. Most options
will be self-explanatory. Note that you can rename a variable and can edit its descriptive label
under “Edit attributes”. You can also “Define a new variable” via a formula (e.g. involving
Chapter 2. Getting started 10
some function of one or more existing variables). For the syntax of such formulae, look at the
online help for “Generate variable syntax” or see the genr command in the Gretl Command
Reference. One simple example:
foo = x1 * x2
will create a new variable foo as the product of the existing variables x1 and x2. In these
formulae, variables must be referenced by name, not number.
• Model menu For details on the various estimators offered under this menu please consult the
Gretl Command Reference. Also see Chapter 22 regarding the estimation of nonlinear models.
• Help menu Please use this as needed! It gives details on the syntax required in various dialog
entries.
Return Opens a window displaying the values of the currently selected variables: it is
the same as selecting “Data, Display Values”.
Delete Pressing this key has the effect of deleting the selected variables. A confirma-
tion is required, to prevent accidental deletions.
e Has the same effect as selecting “Edit attributes” from the “Variable” menu.
F2 Same as “e”. Included for compatibility with other programs.
g Has the same effect as selecting “Define new variable” from the “Variable”
menu (which maps onto the genr command).
h Opens a help window for gretl commands.
F1 Same as “h”. Included for compatibility with other programs.
r Refreshes the variable list in the main window.
t Graphs the selected variable; a line graph is used for time-series datasets,
whereas a distribution plot is used for cross-sectional data.
The icons have the following functions, reading from left to right:
1. Launch a calculator program. A convenience function in case you want quick access to a
calculator when you’re working in gretl. The default program is calc.exe under MS Win-
dows, or xcalc under the X window system. You can change the program under the “Tools,
Preferences, General” menu, “Programs” tab.
2. Start a new script. Opens an editor window in which you can type a series of commands to be
sent to the program as a batch.
3. Open the gretl console. A shortcut to the “Gretl console” menu item (Section 2.3 above).
7. Open the help item for script commands syntax (i.e. a listing with details of all available
commands).
9. Open the dialog box for estimating a model using ordinary least squares.
10. Open a window listing the sample datasets supplied with gretl, and any other data file collec-
tions that have been installed.
Chapter 3
Modes of working
The toolbar at the top of the script window offers the following functions (left to right): (1) Save
the file; (2) Save the file under a specified name; (3) Print the file (this option is not available on all
platforms); (4) Execute the commands in the file; (5) Copy selected text; (6) Paste the selected text;
(7) Find and replace text; (8) Undo the last Paste or Replace action; (9) Help (if you place the cursor
in a command word and press the question mark you will get help on that command); (10) Close
the window.
When you execute the script, by clicking on the Execute icon or by pressing Ctrl-r, all output is
directed to a single window, where it can be edited, saved or copied to the clipboard. To learn
more about the possibilities of scripting, take a look at the gretl Help item “Command reference,”
12
Chapter 3. Modes of working 13
or start up the command-line program gretlcli and consult its help, or consult the Gretl Command
Reference.
If you run the script when part of it is highlighted, gretl will only run that portion. Moreover, if you
want to run just the current line, you can do so by pressing Ctrl-Enter.1
Clicking the right mouse button in the script editor window produces a pop-up menu. This gives
you the option of executing either the line on which the cursor is located, or the selected region of
the script if there’s a selection in place. If the script is editable, this menu also gives the option of
adding or removing comment markers from the start of the line or lines.
The gretl package includes over 70 “practice” scripts. Most of these relate to Ramanathan (2002),
but they may also be used as a free-standing introduction to scripting in gretl and to various points
of econometric theory. You can explore the practice files under “File, Script files, Practice file” There
you will find a listing of the files along with a brief description of the points they illustrate and the
data they employ. Open any file and run it to see the output. Note that long commands in a script
can be broken over two or more lines, using backslash as a continuation character.
You can, if you wish, use the GUI controls and the scripting approach in tandem, exploiting each
method where it offers greater convenience. Here are two suggestions.
• Open a data file in the GUI. Explore the data — generate graphs, run regressions, perform tests.
Then open the Command log, edit out any redundant commands, and save it under a specific
name. Run the script to generate a single file containing a concise record of your work.
• Start by establishing a new script file. Type in any commands that may be required to set
up transformations of the data (see the genr command in the Gretl Command Reference).
Typically this sort of thing can be accomplished more efficiently via commands assembled
with forethought rather than point-and-click. Then save and run the script: the GUI data
window will be updated accordingly. Now you can carry out further exploration of the data
via the GUI. To revisit the data at a later point, open and rerun the “preparatory” script first.
1 This feature is not unique to gretl; other econometric packages offer the same facility. However, experience shows
that while this can be remarkably useful, it can also lead to writing dinosaur scripts that are never meant to be executed
all at once, but rather used as a chaotic repository to cherry-pick snippets from. Since gretl allows you to have several
script windows open at the same time, you may want to keep your scripts tidy and reasonably small.
Chapter 3. Modes of working 14
That is, you type a name for the model to be saved under, then a back-pointing “assignment arrow”,
then the model command. The assignment arrow is composed of the less-than sign followed by a
dash; it must be separated by spaces from both the preceding name and the following command.
The name for a saved object may include spaces, but in that case it must be wrapped in double
quotes:
Models saved in this way will appear as icons in the gretl icon view window (see Section 3.4) after
the script is executed. In addition, you can arrange to have a named model displayed (in its own
window) automatically as follows:
Model1.show
"Model 1".show
The same facility can be used for graphs. For example the following will create a plot of Ct against
Yt, save it under the name “CrossPlot” (it will appear under this name in the icon view window),
and have it displayed:
You can also save the output from selected commands as named pieces of text (again, these will
appear in the session icon window, from where you can open them later). For example this com-
mand sends the output from an augmented Dickey–Fuller test to a “text object” named ADF1 and
displays it in a window:
Objects saved in this way (whether models, graphs or pieces of text output) can be destroyed using
the command .free appended to the name of the object, as in ADF1.free.
In the console, you have “command history”; that is, you can use the up and down arrow keys to
navigate the list of command you have entered to date. You can retrieve, edit and then re-enter a
previous command.
In console mode, you can create, display and free objects (models, graphs or text) aa described
above for script mode.
Figure 3.2: Icon view: one model and one graph have been added to the default icons
If you start gretl and open a data set, then select “Icon view” from the View menu, you should see
the basic default set of icons: these give you quick access to information on the data set (if any),
correlation matrix (“Correlations”) and descriptive summary statistics (“Summary”). All of these
are activated by double-clicking the relevant icon. The “Data set” icon is a little more complex:
double-clicking opens up the data in the built-in spreadsheet, but you can also right-click on the
icon for a menu of other actions.
To add a model to the Icon view, first estimate it using the Model menu. Then pull down the File
menu in the model window and select “Save to session as icon. . . ” or “Save as icon and close”.
Simply hitting the S key over the model window is a shortcut to the latter action.
To add a graph, first create it (under the View menu, “Graph specified vars”, or via one of gretl’s
other graph-generating commands). Click on the graph window to bring up the graph menu, and
select “Save to session as icon”.
Once a model or graph is added its icon will appear in the Icon view window. Double-clicking on the
icon redisplays the object, while right-clicking brings up a menu which lets you display or delete
the object. This popup menu also gives you the option of editing graphs.
table format, therefore gretl will not let those models be added to the model table. These methods
include non-linear least squares (nls), generic maximum-likelihood estimators (mle), generic GMM
(gmm), dynamic panel models (dpanel or its predecessor arbond), interval regressions (intreg),
bivariate probit models (biprobit), AR(I)MA models (arima or arma), and (G)ARCH models (garch
and arch).
In the Icon view window gretl provides a means of constructing such a table (and copying it in plain
text, LATEX or Rich Text Format). The procedure is outlined below. (The model table can also be built
non-interactively, in script mode—see the entry for modeltab in the Gretl Command Reference.)
1. Estimate a model which you wish to include in the table, and in the model display window,
under the File menu, select “Save to session as icon” or “Save as icon and close”.
2. Repeat step 1 for the other models to be included in the table (up to a total of six models).
3. When you are done estimating the models, open the icon view of your gretl session, by se-
lecting “Icon view” under the View menu in the main gretl window, or by clicking the “session
icon view” icon on the gretl toolbar.
4. In the Icon view, there is an icon labeled “Model table”. Decide which model you wish to
appear in the left-most column of the model table and add it to the table, either by dragging
its icon onto the Model table icon, or by right-clicking on the model icon and selecting “Add
to model table” from the pop-up menu.
5. Repeat step 4 for the other models you wish to include in the table. The second model selected
will appear in the second column from the left, and so on.
6. When you are finished composing the model table, display it by double-clicking on its icon.
Under the Edit menu in the window which appears, you have the option of copying the table
to the clipboard in various formats.
7. If the ordering of the models in the table is not what you wanted, right-click on the model
table icon and select “Clear table”. Then go back to step 4 above and try again.
• Start gretl then re-open the session file by going to the “File, Session files, Open session”, or
• From the command line, type gretl -r sessionfile, where sessionfile is the name under which
the session was saved, or
Data files
• Plain text files (comma-separated or “CSV” being the most common type). For details on what
gretl expects of such files, see Section 4.3.
• Spreadsheets: MS Excel, Gnumeric and Open Document (ODS). The requirements for such files
are given in Section 4.3.
When you import data from a plain text format, gretl opens a “diagnostic” window, reporting on its
progress in reading the data. If you encounter a problem with ill-formatted data, the messages in
this window should give you a handle on fixing the problem.
Note that gretl has a facility for writing out data in the native formats of GNU R, Octave, JMulTi and
PcGive (see Appendix E). In the GUI client this option is found under the “File, Export data” menu;
in the command-line client use the store command with the appropriate option flag.
4.2 Databases
For working with large amounts of data gretl is supplied with a database-handling routine. A
database, as opposed to a data file, is not read directly into the program’s workspace. A database
can contain series of mixed frequencies and sample ranges. You open the database and select
series to import into the working dataset. You can then save those series in a native format data
file if you wish. Databases can be accessed via the menu item “File, Databases”.
For details on the format of gretl databases, see Appendix A.
1 See http://ricardo.ecn.wfu.edu/~cottrell/eviews_format/.
18
Chapter 4. Data files 19
+ Visit the gretl data page for details and updates on available data.
1. Find, or create using a text editor, a plain text data file and open it via “Import”.
2. Use your favorite spreadsheet to establish the data file, save it in comma-separated format if
necessary (this may not be necessary if the spreadsheet format is MS Excel, Gnumeric or Open
Document), then use one of the “Import” options.
5. Use your favorite text editor or other software tools to a create data file in gretl format inde-
pendently.
• The first row must contain valid variable names. A valid variable name is of 31 characters
maximum; starts with a letter; and contains nothing but letters, numbers and the underscore
character, _. (Longer variable names will be truncated to 31 characters.) Qualifications to the
above: First, in the case of an plain text import, if the file contains no row with variable names
the program will automatically add names, v1, v2 and so on. Second, by “the first row” is
meant the first relevant row. In the case of plain text imports, blank rows and rows beginning
with a hash mark, #, are ignored. In the case of Excel, Gnumeric and ODS imports, you are
presented with a dialog box where you can select an offset into the spreadsheet, so that gretl
will ignore a specified number of rows and/or columns.
• Data values: these should constitute a rectangular block, with one variable per column (and
one observation per row). The number of variables (data columns) must match the number
of variable names given. See also section 4.6. Numeric data are expected, but in the case of
Chapter 4. Data files 20
importing from plain text, the program offers limited handling of character (string) data: if
a given column contains character data only, consecutive numeric codes are substituted for
the strings, and once the import is complete a table is printed showing the correspondence
between the strings and the codes.
• Dates (or observation labels): Optionally, the first column may contain strings such as dates,
or labels for cross-sectional observations. Such strings have a maximum of 15 characters (as
with variable names, longer strings will be truncated). A column of this sort should be headed
with the string obs or date, or the first row entry may be left blank.
For dates to be recognized as such, the date strings should adhere to one or other of a set of
specific formats, as follows. For annual data: 4-digit years. For quarterly data: a 4-digit year,
followed by a separator (either a period, a colon, or the letter Q), followed by a 1-digit quarter.
Examples: 1997.1, 2002:3, 1947Q1. For monthly data: a 4-digit year, followed by a period or
a colon, followed by a two-digit month. Examples: 1997.01, 2002:10.
Plain text (“CSV”) files can use comma, space, tab or semicolon as the column separator. When you
open such a file via the GUI you are given the option of specifying the separator, though in most
cases it should be detected automatically.
If you use a spreadsheet to prepare your data you are able to carry out various transformations of
the “raw” data with ease (adding things up, taking percentages or whatever): note, however, that
you can also do this sort of thing easily — perhaps more easily — within gretl, by using the tools
under the “Add” menu.
opens if you choose the “Gretl native” option since this file, which contains a large collection of US
macroeconomic time series, is supplied with the distribution.
You won’t find anything under “RATS 4” unless you have purchased RATS data.2 If you do possess
RATS data you should go into the “Tools, Preferences, General” dialog, select the Databases tab,
and fill in the correct path to your RATS files.
If your computer is connected to the internet you should find several databases (at Wake Forest
University) under “On database server”. You can browse these remotely; you also have the option
of installing them onto your own computer. The initial remote databases window has an item
showing, for each file, whether it is already installed locally (and if so, if the local version is up to
date with the version at Wake Forest).
Assuming you have managed to open a database you can import selected series into gretl’s workspace
by using the “Series, Import” menu item in the database window, or via the popup menu that ap-
pears if you click the right mouse button, or by dragging the series into the program’s main window.
1. Cross section
2. Time series
3. Panel data
The primary tool for doing this is the “Data, Dataset structure” menu entry in the graphical inter-
face, or the setobs command for scripts and the command-line interface.
2 See www.estima.com
Chapter 4. Data files 22
and click “Forward” once more. Finally, click “OK” to confirm the time-series interpretation if it is
correct (or click “Back” to make adjustments if need be).
Besides the basic business of getting a data set interpreted as time series, further issues may arise
relating to the frequency of time-series data. In a gretl time-series data set, all the series must
have the same frequency. Suppose you wish to make a combined dataset using series that, in their
original state, are not all of the same frequency. For example, some series are monthly and some
are quarterly.
Your first step is to formulate a strategy: Do you want to end up with a quarterly or a monthly data
set? A basic point to note here is that “compacting” data from a higher frequency (e.g. monthly) to
a lower frequency (e.g. quarterly) is usually unproblematic. You lose information in doing so, but
in general it is perfectly legitimate to take (say) the average of three monthly observations to create
a quarterly observation. On the other hand, “expanding” data from a lower to a higher frequency is
not, in general, a valid operation.
In most cases, then, the best strategy is to start by creating a data set of the lower frequency, and
then to compact the higher frequency data to match. When you import higher-frequency data from
a database into the current data set, you are given a choice of compaction method (average, sum,
start of period, or end of period). In most instances “average” is likely to be appropriate.
You can also import lower-frequency data into a high-frequency data set, but this is generally not
recommended. What gretl does in this case is simply replicate the values of the lower-frequency
series as many times as required. For example, suppose we have a quarterly series with the value
35.5 in 1990:1, the first quarter of 1990. On expansion to monthly, the value 35.5 will be assigned
to the observations for January, February and March of 1990. The expanded variable is therefore
useless for fine-grained time-series analysis, outside of the special case where you know that the
variable in question does in fact remain constant over the sub-periods.
When the current data frequency is appropriate, gretl offers both “Compact data” and “Expand
data” options under the “Data” menu. These options operate on the whole data set, compacting or
exanding all series. They should be considered “expert” options and should be used with caution.
Panel data
Panel data are inherently three dimensional — the dimensions being variable, cross-sectional unit,
and time-period. For example, a particular number in a panel data set might be identified as the
observation on capital stock for General Motors in 1980. (A note on terminology: we use the
terms “cross-sectional unit”, “unit” and “group” interchangeably below to refer to the entities that
compose the cross-sectional dimension of the panel. These might, for instance, be firms, countries
or persons.)
For representation in a textual computer file (and also for gretl’s internal calculations) the three
dimensions must somehow be flattened into two. This “flattening” involves taking layers of the
data that would naturally stack in a third dimension, and stacking them in the vertical dimension.
gretl always expects data to be arranged “by observation”, that is, such that each row represents an
observation (and each variable occupies one and only one column). In this context the flattening of
a panel data set can be done in either of two ways:
• Stacked time series: the successive vertical blocks each comprise a time series for a given
unit.
• Stacked cross sections: the successive vertical blocks each comprise a cross-section for a
given period.
You may input data in whichever arrangement is more convenient. Internally, however, gretl always
stores panel data in the form of stacked time series.
Chapter 4. Data files 23
where unitvar is a variable that indexes the units and timevar is a variable indexing the periods.
Alternatively you can use the form setobs freq 1:1 structure, where freq is replaced by the “block
size” of the data (that is, the number of periods in the case of stacked time series, or the number
of units in the case of stacked cross-sections) and structure is either --stacked-time-series or
--stacked-cross-section. Two examples are given below: the first is suitable for a panel in
the form of stacked time series with observations from 20 periods; the second for stacked cross
sections with 5 units.
x1
1965 1970 1975 1980 1985
AR 100.0 110.5 118.7 131.2 160.4
AZ 100.0 104.3 113.8 120.9 140.6
If a datafile with this sort of structure is read into gretl,3 the program will interpret the columns as
distinct variables, so the data will not be usable “as is.” But there is a mechanism for correcting the
situation, namely the stack function.
Consider the first data column in the fragment above: the first 50 rows of this column constitute a
cross-section for the variable x1 in the year 1965. If we could create a new series by stacking the
3 Note that you will have to modify such a datafile slightly before it can be read at all. The line containing the variable
name (in this example x1) will have to be removed, and so will the initial row containing the years, otherwise they will be
taken as numerical data.
Chapter 4. Data files 24
first 50 entries in the second column underneath the first 50 entries in the first, we would be on the
way to making a data set “by observation” (in the first of the two forms mentioned above, stacked
cross-sections). That is, we’d have a column comprising a cross-section for x1 in 1965, followed by
a cross-section for the same variable in 1970.
The following gretl script illustrates how we can accomplish the stacking, for both x1 and x2. We
assume that the original data file is called panel.txt, and that in this file the columns are headed
with “variable names” v1, v2, . . . , v5. (The columns are not really variables, but in the first instance
we “pretend” that they are.)
open panel.txt
series x1 = stack(v1..v5) --length=50
series x2 = stack(v1..v5) --offset=50 --length=50
setobs 50 1:1 --stacked-cross-section
store panel.gdt x1 x2
The second line illustrates the syntax of the stack function. The double dots within the parenthe-
ses indicate a range of variables to be stacked: here we want to stack all 5 columns (for all 5 years).4
The full data set contains 100 rows; in the stacking of variable x1 we wish to read only the first 50
rows from each column: we achieve this by adding --length=50. Note that if you want to stack a
non-contiguous set of columns you can give a comma-separated list of variable names, as in
series x = stack(v1,v3,v5)
or you can provide within the parentheses the name of a previously created list (see chapter 14).
On line 3 we do the stacking for variable x2. Again we want a length of 50 for the components of
the stacked series, but this time we want gretl to start reading from the 50th row of the original
data, and we specify --offset=50. Line 4 imposes a panel interpretation on the data; finally, we
save the data in gretl format, with the panel interpretation, discarding the original “variables” v1
through v5.
The illustrative script above is appropriate when the number of variables to be processed is small.
When then are many variables in the data set it will be more efficient to use a loop to accomplish
the stacking, as shown in the following script. The setup is presumed to be the same as in the
previous section (50 units, 5 periods), but with 20 variables rather than 2.
open panel.txt
loop i=1..20
scalar k = (i - 1) * 50
series x$i = stack(v1..v5) --offset=k --length=50
endloop
setobs 50 1.01 --stacked-cross-section
store panel.gdt x1 x2 x3 x4 x5 x6 x7 x8 x9 x10 \
x11 x12 x13 x14 x15 x16 x17 x18 x19 x20
genr time
series year = 1960 + (5 * time)
genr markers = "%s:%d", marker, year
The first line generates a 1-based index representing the period of each observation, and the second
line uses the time variable to generate a variable representing the year of the observation. The
third line contains this special feature: if (and only if) the name of the new “variable” to generate is
markers, the portion of the command following the equals sign is taken as a C-style format string
(which must be wrapped in double quotes), followed by a comma-separated list of arguments.
The arguments will be printed according to the given format to create a new set of observation
markers. Valid arguments are either the names of variables in the dataset, or the string marker
which denotes the pre-existing observation marker. The format specifiers which are likely to be
useful in this context are %s for a string and %d for an integer. Strings can be truncated: for
example %.3s will use just the first three characters of the string. To chop initial characters off
an existing observation marker when constructing a new one, you can use the syntax marker + n,
where n is a positive integer: in the case the first n characters will be skipped.
After the commands above are processed, then, the observation markers will look like, for example,
AR:1965, where the two-letter state code and the year of the observation are spliced together with
a colon.
1. “unit dummies” (script command genr unitdum). This command creates a set of dummy
variables identifying the cross-sectional units. The variable du_1 will have value 1 in each
row corresponding to a unit 1 observation, 0 otherwise; du_2 will have value 1 in each row
corresponding to a unit 2 observation, 0 otherwise; and so on.
2. “time dummies” (script command genr timedum). This command creates a set of dummy
variables identifying the periods. The variable dt_1 will have value 1 in each row correspond-
ing to a period 1 observation, 0 otherwise; dt_2 will have value 1 in each row corresponding
to a period 2 observation, 0 otherwise; and so on.
If a panel data set has the YEAR of the observation entered as one of the variables you can create a
periodic dummy to pick out a particular year, e.g. genr dum = (YEAR==1960). You can also create
periodic dummy variables using the modulus operator, %. For instance, to create a dummy with
value 1 for the first observation and every thirtieth observation thereafter, 0 otherwise, do
genr index
series dum = ((index-1) % 30) == 0
of x1 where available, and the missing value code where the lag is not available (e.g. at the start of
the time series for each group). When you run a regression using such variables, the program will
automatically skip the missing observations.
When a panel data set has a fairly substantial time dimension, you may wish to include a trend in
the analysis. The command genr time creates a variable named time which runs from 1 to T for
each unit, where T is the length of the time-series dimension of the panel. If you want to create an
index that runs consecutively from 1 to m × T , where m is the number of units in the panel, use
genr index.
creates a series of this form: the first 8 values (corresponding to unit 1) contain the mean of x for
unit 1, the next 8 values contain the mean for unit 2, and so on. The psd() function works in a
similar manner. The sample standard deviation for group i is computed as
sP
(x − x̄i )2
si =
Ti − 1
where Ti denotes the number of valid observations on x for the given unit, x̄i denotes the group
mean, and the summation is across valid observations for the group. If Ti < 2, however, the
standard deviation is recorded as 0.
One particular use of psd() may be worth noting. If you want to form a sub-sample of a panel that
contains only those units for which the variable x is time-varying, you can either use
or
• In calculating descriptive statistics (mean, standard deviation, etc.) under the summary com-
mand, missing values are simply skipped and the sample size adjusted appropriately.
• In running regressions gretl first adjusts the beginning and end of the sample range, trun-
cating the sample if need be. Missing values at the beginning of the sample are common in
Chapter 4. Data files 27
time series work due to the inclusion of lags, first differences and so on; missing values at the
end of the range are not uncommon due to differential updating of series and possibly the
inclusion of leads.
If gretl detects any missing values “inside” the (possibly truncated) sample range for a regression,
the result depends on the character of the dataset and the estimator chosen. In many cases, the
program will automatically skip the missing observations when calculating the regression results.
In this situation a message is printed stating how many observations were dropped. On the other
hand, the skipping of missing observations is not supported for all procedures: exceptions include
all autoregressive estimators, system estimators such as SUR, and nonlinear least squares. In the
case of panel data, the skipping of missing observations is supported only if their omission leaves
a balanced panel. If missing observations are found in cases where they are not supported, gretl
gives an error message and refuses to produce estimates.
The function zeromiss(), which again takes a single series as its argument, returns a series where
all zero values are set to the missing code. This should be used with caution — one does not want
to confuse missing values and zeros— but it can be useful in some contexts. For example, one can
determine the first valid observation for a variable x using
genr time
scalar x0 = min(zeromiss(time * ok(x)))
The function misszero() does the opposite of zeromiss, that is, it converts all missing values to
zero.
If missing values get involved in calculations, they propagate according to the IEEE rules: notably,
if one of the operands to an arithmetical operation is a NaN, the result will also be NaN.
If RAM is not an issue, there is one further limitation on data size (though it’s very unlikely to
be a binding constraint). That is, variables and observations are indexed by signed integers, and
on a typical PC these will be 32-bit values, capable of representing a maximum positive value of
231 − 1 = 2, 147, 483, 647.
The limits mentioned above apply to gretl’s “native” functionality. There are tighter limits with
regard to two third-party programs that are available as add-ons to gretl for certain sorts of time-
series analysis including seasonal adjustment, namely TRAMO/SEATS and X-12-ARIMA. These pro-
grams employ a fixed-size memory allocation, and can’t handle series of more than 600 observa-
tions.
• For data files: select the menu item “File, Open data, Sample file”, or click on the folder icon
on the gretl toolbar.
• For script files: select the menu item “File, Script files, Practice file”.
• The data or script files included in the gretl distribution are automatically shown (this includes
files relating to Ramanathan’s Introductory Econometrics and Greene’s Econometric Analysis).
• The program looks for certain known collections of data files available as optional extras,
for instance the datafiles from various econometrics textbooks (Davidson and MacKinnon,
Gujarati, Stock and Watson, Verbeek, Wooldridge) and the Penn World Table (PWT 5.6). (See
the data page at the gretl website for information on these collections.) If the additional files
are found, they are added to the selection windows.
• The program then searches for valid file collections (not necessarily known in advance) in
these places: the “system” data directory, the system script directory, the user directory,
and all first-level subdirectories of these. For reference, typical values for these directories
are shown in Table 4.1. (Note that PERSONAL is a placeholder that is expanded by Windows,
corresponding to “My Documents” on English-language systems.)
Linux MS Windows
system data dir /usr/share/gretl/data c:\Program Files\gretl\data
system script dir /usr/share/gretl/scripts c:\Program Files\gretl\scripts
user dir $HOME/gretl PERSONAL\gretl
Any valid collections will be added to the selection windows. So what constitutes a valid file collec-
tion? This comprises either a set of data files in gretl XML format (with the .gdt suffix) or a set of
script files containing gretl commands (with .inp suffix), in each case accompanied by a “master
file” or catalog. The gretl distribution contains several example catalog files, for instance the file
descriptions in the misc sub-directory of the gretl data directory and ps_descriptions in the
misc sub-directory of the scripts directory.
Chapter 4. Data files 29
If you are adding your own collection, data catalogs should be named descriptions and script
catalogs should be be named ps_descriptions. In each case the catalog should be placed (along
with the associated data or script files) in its own specific sub-directory (e.g. /usr/share/gretl/
data/mydata or c:\userdata\gretl\data\mydata).
The catalog files are plain text; if they contain non-ASCII characters they must be encoded as UTF-
8. The syntax of such files is straightforward. Here, for example, are the first few lines of gretl’s
“misc” data catalog:
The first line, which must start with a hash mark, contains a short name, here “Gretl”, which
will appear as the label for this collection’s tab in the data browser window, followed by a colon,
followed by an optional short description of the collection.
Subsequent lines contain two elements, separated by a comma and wrapped in double quotation
marks. The first is a datafile name (leave off the .gdt suffix here) and the second is a short de-
scription of the content of that datafile. There should be one such line for each datafile in the
collection.
A script catalog file looks very similar, except that there are three fields in the file lines: a filename
(without its .inp suffix), a brief description of the econometric point illustrated in the script, and
a brief indication of the nature of the data used. Again, here are the first few lines of the supplied
“misc” script catalog:
If you want to make your own data collection available to users, these are the steps:
2. Convert the data to gretl format and save as gdt files. It is probably easiest to convert the data
by importing them into the program from plain text, CSV, or a spreadsheet format (MS Excel
or Gnumeric) then saving them. You may wish to add descriptions of the individual variables
(the “Variable, Edit attributes” menu item), and add information on the source of the data (the
“Data, Edit info” menu item).
4. Put the datafiles plus the descriptions file in a subdirectory of the gretl data directory (or user
directory).
5. If the collection is to be distributed to other people, package the data files and catalog in some
suitable manner, e.g. as a zipfile.
If you assemble such a collection, and the data are not proprietary, we would encourage you to
submit the collection for packaging as a gretl optional extra.
Chapter 4. Data files 30
Sub-sampling a dataset
5.1 Introduction
Some subtle issues can arise here; this chapter attempts to explain the issues.
A sub-sample may be defined in relation to a full dataset in two different ways: we will refer to these
as “setting” the sample and “restricting” the sample; these methods are discussed in sections 5.2
and 5.3 respectively. In addition section 5.4 discusses some special issues relating to panel data,
and section 5.5 covers resampling with replacement, which is useful in the context of bootstrapping
test statistics.
The following discussion focuses on the command-line approach. But you can also invoke the
methods outlined here via the items under the Sample menu in the GUI program.
Or one wishes to set aside a block of observations at the end of the data period for out-of-sample
forecasting. In that case one might do
smpl ; 2000:4
where the semicolon is shorthand for “leave the starting observation unchanged”. (The semicolon
may also be used in place of the second parameter, to mean that the ending observation should be
unchanged.) By “unchanged” here, we mean unchanged relative to the last smpl setting, or relative
to the full dataset if no sub-sample has been defined up to this point. For example, after
smpl +1 ;
will advance the starting observation by one while preserving the ending observation, and
smpl +2 -1
31
Chapter 5. Sub-sampling a dataset 32
will both advance the starting observation by two and retard the ending observation by one.
An important feature of “setting” the sample as described above is that it necessarily results in
the selection of a subset of observations that are contiguous in the full dataset. The structure of
the dataset is therefore unaffected (for example, if it is a quarterly time series before setting the
sample, it remains a quarterly time series afterwards).
to this effect. Or suppose we want to restrict the sample to respondents with incomes over $50,000.
Then we could use
A question arises: if we issue the two commands above in sequence, what do we end up with in
our sub-sample: all cases with income over 50000, or just women with income over 50000? By
default, the answer is the latter: women with income over 50000. The second restriction augments
the first, or in other words the final restriction is the logical product of the new restriction and any
restriction that is already in place. If you want a new restriction to replace any existing restrictions
you can first recreate the full dataset using
smpl --full
Alternatively, you can add the replace option to the smpl command:
This option has the effect of automatically re-establishing the full dataset before applying the new
restriction.
Unlike a simple “setting” of the sample, “restricting” the sample may result in selection of non-
contiguous observations from the full data set. It may therefore change the structure of the data
set.
This can be seen in the case of panel data. Say we have a panel of five firms (indexed by the variable
firm) observed in each of several years (identified by the variable year). Then the restriction
produces a dataset that is not a panel, but a cross-section for the year 1995. Similarly
the case of setting, the program merely records the starting and ending observations and uses these
as parameters to the various commands calling for the estimation of models, the computation of
statistics, and so on. In the case of restriction, the program makes a reduced copy of the dataset
and by default treats this reduced copy as a simple, undated cross-section — but see the further
discussion of panel data in section 5.4.
If you wish to re-impose a time-series interpretation of the reduced dataset you can do so using the
setobs command, or the GUI menu item “Data, Dataset structure”.
The fact that “restricting” the sample results in the creation of a reduced copy of the original
dataset may raise an issue when the dataset is very large. With such a dataset in memory, the
creation of a copy may lead to a situation where the computer runs low on memory for calculating
regression results. You can work around this as follows:
1. Open the full data set, and impose the sample restriction.
2. Save a copy of the reduced data set to disk.
3. Close the full dataset and open the reduced one.
4. Proceed with your analysis.
Random sub-sampling
Besides restricting the sample on some deterministic criterion, it may sometimes be useful (when
working with very large datasets, or perhaps to study the properties of an estimator) to draw a
random sub-sample from the full dataset. This can be done using, for example,
to select 100 cases. If you want the sample to be reproducible, you should set the seed for the
random number generator first, using the set command. This sort of sampling falls under the
“restriction” category: a reduced copy of the dataset is made.
is to limit the sample to the first 80 firms. Note that if you instead tried smpl 1:1 80:4 this would
provoke an error: you cannot use this syntax to sub-sample in the time dimension of the panel.
Alternatively, and perhaps more naturally, you can use the --unit option with the smpl command
to limit the sample in the cross-sectional dimension, as in
smpl 1 80 --unit
The firms in the Arellano–Bond dataset are anonymous, but suppose you had a panel with five
named countries. With such a panel you can inform gretl of the names of the groups using the
setobs command. For example, given
Chapter 5. Sub-sampling a dataset 34
gretl creates a string-valued series named country with group names taken from the variable cstr.
Then, to include only Italy and Spain you could do
To sub-sample in the time dimension, use of --restrict is required. For example, the Arellano–
Bond dataset contains a variable named YEAR that records the year of the observations and if one
wanted to omit the first two years of data one could do
If a dataset does not already incude a suitable variable for this purpose one can use the command
genr time to create a simple 1-based time index.
Note that if you apply a sample restriction that just selects certain units (firms, countries or what-
ever), or selects certain contiguous time-periods — such that n > 1, T > 1 and the time-series length
is still the same across all included units — your sub-sample will still be interpreted by gretl as a
panel.
Unbalancing restrictions
In some cases one wants to sub-sample according to a criterion that “cuts across the grain” of
a panel dataset. For instance, suppose you have a micro dataset with thousands of individuals
observed over several years and you want to restrict the sample to observations on employed
women.
If we simply extracted from the total nT rows of the dataset those that pertain to women who were
employed at time t (t = 1, . . . , T ) we would likely end up with a dataset that doesn’t count as a
panel in gretl (because the specific time-series length, Ti , would differ across individuals). In some
contexts it might be OK that gretl doesn’t take your sub-sample to be a panel, but if you want to
apply panel-specific methods this is a problem. You can solve it by giving the --balanced option
with smpl. For example, supposing your dataset contained dummy variables gender (with the value
1 coding for women) and employed, you could do
What exactly does this do? Well, let’s say the years of your data are 2000, 2005 and 2010, and that
some women were employed in all of those years, giving a maximum Ti value of 3. But individual
526 is a women who was employed only in the year 2000 (Ti = 1). The effect of the --balanced
option is then to insert “padding rows” of NAs for the years 2005 and 2010 for individual 526, and
similarly for all individuals with 0 < Ti < 3. Your sub-sample then qualifies as a panel.
series xr = resample(x)
Chapter 5. Sub-sampling a dataset 35
creates a new series each of whose elements is drawn at random from the elements of x. If the
original series has 100 observations, each element of x is selected with probability 1/100 at each
drawing. Thus the effect is to “shuffle” the elements of x, with the twist that each element of x may
appear more than once, or not at all, in xr.
The primary use of this function is in the construction of bootstrap confidence intervals or p-values.
Here is a simple example. Suppose we estimate a simple regression of y on x via OLS and find that
the slope coefficient has a reported t-ratio of 2.5 with 40 degrees of freedom. The two-tailed p-
value for the null hypothesis that the slope parameter equals zero is then 0.0166, using the t(40)
distribution. Depending on the context, however, we may doubt whether the ratio of coefficient to
standard error truly follows the t(40) distribution. In that case we could derive a bootstrap p-value
as shown in Listing 5.1.
Under the null hypothesis that the slope with respect to x is zero, y is simply equal to its mean plus
an error term. We simulate y by resampling the residuals from the initial OLS and re-estimate the
model. We repeat this procedure a large number of times, and record the number of cases where
the absolute value of the t-ratio is greater than 2.5: the proportion of such cases is our bootstrap
p-value. For a good discussion of simulation-based tests and bootstrapping, see Davidson and
MacKinnon (2004, chapter 4); Davidson and Flachaire (2001) is also instructive.
ols y 0 x
# save the residuals
genr ui = $uhat
scalar ybar = mean(y)
# number of replications for bootstrap
scalar replics = 10000
scalar tcount = 0
series ysim
loop replics --quiet
# generate simulated y by resampling
ysim = ybar + resample(ui)
ols ysim 0 x
scalar tsim = abs($coeff(x) / $stderr(x))
tcount += (tsim > 2.5)
endloop
printf "proportion of cases with |t| > 2.5 = %g\n", tcount / replics
Chapter 6
• Save as PNG: Save the graph in Portable Network Graphics format (the same format that you
see on screen).
• Save to session as icon: The graph will appear in iconic form when you select “Icon view” from
the View menu.
• Zoom: Lets you select an area within the graph for closer inspection (not available for all
graphs).
• Print: (Current GTK or MS Windows only) lets you print the graph directly.
• Copy to clipboard: MS Windows only, lets you paste the graph into Windows applications such
as MS Word.
• Edit: Opens a controller for the plot which lets you adjust many aspects of its appearance.
1 For an example of such a dataset, see the Ramanathan file data4-10: this contains data on private school enrollment
for the 50 states of the USA plus Washington, DC; the case markers are the two-letter codes for the states.
36
Chapter 6. Graphs and plots 37
• In the graph display window, right-click and choose “Save to session as icon”.
• If it’s not already open, open the icon view window — either via the menu item View/Icon view,
or by clicking the “session icon view” button on the main-window toolbar.
• Right-click on the icon representing the newly added graph and select “Edit plot commands”
from the pop-up menu.
Here are the basic things you can do in this window. Obviously, you can edit the file you just
opened. You can also send it for processing by gnuplot, by clicking the “Execute” (cogwheel) icon
in the toolbar. Or you can use the “Save as” button to save a copy for editing and processing as you
wish.
Chapter 6. Graphs and plots 38
Unless you’re a gnuplot expert, most likely you’ll only need to edit a couple of lines at the top of
the file, specifying a driver (plus options) and an output file. We offer here a brief summary of some
points that may be useful.
First, gnuplot’s output mode is set via the command set term followed by the name of a supported
driver (“terminal” in gnuplot parlance) plus various possible options. (The top line in the plot
commands window shows the set term line that gretl used to make a PNG file, commented out.)
The graphic formats that are most suitable for publication are PDF and EPS. These are supported
by the gnuplot term types pdf, pdfcairo and postscript (with the eps option). The pdfcairo
driver has the virtue that is behaves in a very similar manner to the PNG one, the output of which
you see on screen. This is provided by the version of gnuplot that is included in the gretl packages
for MS Windows and Mac OS X; if you’re on Linux it may or may be supported. If pdfcairo is not
available, the pdf terminal may be available; the postscript terminal is almost certainly available.
Besides selecting a term type, if you want to get gnuplot to write the actual output file you need
to append a set output line giving a filename. Here are a few examples of the first two lines you
might type in the window editing your plot commands. We’ll make these more “realistic” shortly.
There are a couple of things worth remarking here. First, you may want to adjust the size of the
graph, and second you may want to change the font. The default sizes produced by the above
drivers are 5 inches by 3 inches for pdfcairo and pdf, and 5 inches by 3.5 inches for postscript
eps. In each case you can change this by giving a size specification, which takes the form XX,YY
(examples below).
Chapter 6. Graphs and plots 39
You may ask, why bother changing the size in the gnuplot command file? After all, PDF and EPS are
both vector formats, so the graphs can be scaled at will. True, but a uniform scaling will also affect
the font size, which may end looking wrong. You can get optimal results by experimenting with
the font and size options to gnuplot’s set term command. Here are some examples (comments
follow below).
# postscript, regular
set term post eps solid font "Helvetica,16"
# or small
set term post eps solid font "Helvetica,12" size 3in,2in
On the first line we set a sans serif font for pdfcairo at a suitable size for a 5 × 3.5 inch plot
(which you may find looks better than the rather “letterboxy” default of 5 × 3). And on the second
we illustrate what you might do to get a smaller 3 × 2 inch plot. You can specify the plot size in
centimeters if you prefer, as in
We then repeat the exercise for the pdf terminal. Notice that here we’re specifying one of the 35
standard PostScript fonts, namely Helvetica. Unlike pdfcairo, the plain pdf driver is unlikely to
be able to find fonts other than these.
In the third pair of lines we illustrate options for the postscript driver (which, as you see, can
be abbreviated as post). Note that here we have added the option solid. Unlike most other
drivers, this one uses dashed lines unless you specify the solid option. Also note that we’ve
(apparently) specified a much larger font in this case. That’s because the eps option in effect tells
the postscript driver to work at half-size (among other things), so we need to double the font
size.
Table 6.1 summarizes the basics for the three drivers we have mentioned.
To find out more about gnuplot visit www.gnuplot.info. This site has documentation for the current
version of the program in various formats.
Additional tips
To be written. Line widths, enhanced text. Show a “before and after” example.
Chapter 6. Graphs and plots 40
gnuplot y1 y2 x
will give you a basic XY plot of the two series y1 and y2 on the vertical axis versus the series x on
the horizontal axis. In general, the arguments to the gnuplot command is a list of series, the last
of which goes on the x-axis, while all the other ones go onto the y-axis. By default, the gnuplot
command gives you a scatterplot. If you just have one variable on the y-axis, then gretl will also
draw a the OLS interpolation, if the fit is good enough.2
Several aspects of the behavior described above can be modified. You do this by appending options
to the command. Most options can be broadly grouped in three categories:
1. Plot styles: we support points (the default choice), lines, lines and points together, and im-
pulses (vertical lines).
2. Algorithm for the fitted line: here you can choose between linear, quadratic and cubic inter-
polation, but also more exotic choices, such as semi-log, inverse or loess (non-parametric). Of
course, you can also turn this feature off.
3. Input and output: you can choose whether you want your graph on your computer screen
(and possibly use the in-built graphical widget to further customize it — see above, page 37),
or rather save it to a file. We support several graphical formats, among which PNG and PDF,
to make it easy to incorporate your plots into text documents.
The following script uses the AWM dataset to exemplify some traditional plots in macroeconomics:
2 The technical condition for this is that the two-tailed p-value for the slope coefficient should be under 10%.
Chapter 6. Graphs and plots 41
options: specify multiple options on a single line; if more than one option is given on a line, the
options should be separated by spaces.
printf: a printf statement whose result will be passed to gnuplot literally; this allows the use of
string variables without having to resort to @-style string substitution.
The options available are basically those of the current gnuplot command, but with a few dif-
ferences. For one thing you don’t need the leading double-dash in an "option" (or "options") line.
Besides that,
• You can’t use the option --matrix=whatever with plot: that possibility is handled by pro-
viding the name of a matrix on the initial plot line.
• The --input=filename option is not supported: use gnuplot for the case where you’re
supplying the entire plot specification yourself.
• The several options pertaining to the presence and type of a fitted line, are replaced in plot
by a single option fit which requires a parameter. Supported values for the parameter are:
none, linear, quadratic, cubic, inverse, semilog and loess. Example:
option fit=quadratic
As with gnuplot, the default is to show a linear fit in an X-Y scatter if it’s significant at the 10
percent level.
Here’s a simple example, the plot specification from the “bandplot” package, which shows how
to achieve the same result via the gnuplot command and a plot block, respectively — the latter
occupies a few more lines but is clearer
plot plotmat
options with-lines fit=none
literal set linetype 3 lc rgb "#0000ff"
literal set nokey
printf "set title \"%s\"", title
printf "set xlabel \"%s\"", xname
end plot --output=display
Note that --output=display is appended to end plot; also note that if you give a matrix to plot
it’s assumed you want to plot all the columns. In addition, if you give a single series and the dataset
is time series, it’s assumed you want a time-series plot.
Script 6.1 contains a slightly more elaborate example: here we load the Mroz example dataset and
calculate the log of the individual’s wage. Then, we match the histogram of a discretized version
of the same variable (obtained via the aggregate() function) versus the theoretical density if data
were Gaussian.
There are a few points to note:
• The data for the plot are passed through a matrix in which we set column names via the
cnameset function; those names are then automatically used by the plot environment.
• In this example, we make extensive use of the set literal construct for refining the plot by
passing instruction to gnuplot; the power of gnuplot is impossible to overstate. We encourage
you to visit the “demos” version of gnuplot’s website (http://gnuplot.sourceforge.net/)
and revel in amazement.
• In the plot environment you can use all the quantities you have in your script. This is the
way we calibrate the histogram width (try setting the scalar k in the script to different values).
Note that the printf command has a special meaning inside a plot environment.
• The script displays the plot on your screen. If you want to save it to a file instead, replace
--output=display at the end with --output=filename.
• It’s OK to insert comments in the plot environment; actually, it’s a rather good idea to com-
ment as much as possible (as always)!
Listing 6.1: Plotting the log wage from the Mroz example dataset
###
### prepare matrix with data for plot
###
###
### create plot
###
plot plotmat
# move legend
literal set key outside rmargin
# set line style
literal set linetype 2 dashtype 2 linewidth 2
# set histogram color
literal set linetype 1 lc rgb "#777777"
# set histogram style
literal set style fill solid 0.25 border
# set histogram width
printf "set boxwidth %4.2f\n", 0.5/k
options with-lines=2 with-boxes=1
end plot --output=display
Chapter 6. Graphs and plots 44
0.18
frequency
density
0.16
0.14
0.12
0.1
0.08
0.06
0.04
0.02
0
-2 -1 0 1 2 3
log wage
6.3 Boxplots
These plots (after Tukey and Chambers) display the distribution of a variable. Its shape depends
on a few quantities, defined as follows:
The central box encloses the middle 50 percent of the data, i.e. goes from Q1 to Q3 ; therefore, its
height equals R. A line is drawn across the box at the median m and a “+” sign identifies the mean
x̄.
The length of the “whiskers” depends on the presence of outliers. The top whisker extends from
the top of the box up to a maximum of 1.5 times the interquartile range, but can be shorter if the
sample maximum is lower than that value; that is, it reaches min[xmax , Q3 + 1.5R]. Observations
larger than Q3 + 1.5R, if any, are considered outliers and represented individually via dots.3 The
bottom whisker obeys the same logic, with obvious adjustments. Figure 6.4 provides an example
of all this by using the variable FAMINC from the sample dataset mroz87.
In the case of boxplots with confidence intervals, dotted lines show the limits of an approximate 90
percent confidence interval for the median. This is obtained by the bootstrap method, which can
take a while if the data series is very long. For details on constructing boxplots, see the entry for
3 To give you an intuitive idea, if a variable is normally distributed, the chances of picking an outlier by this definition
FAMINC
xmax
80000
outliers
60000
40000
Q3 x̄
20000 m
Q1
0 xmin
boxplot in the Gretl Command Reference or use the Help button that appears when you select one
of the boxplot items under the menu item “View, Graph specified vars” in the main gretl window.
Factorized boxplots
A nice feature which is quite useful for data visualization is the conditional, or factorized boxplot.
This type of plot allows you to examine the distribution of a variable conditional on the value of
some discrete factor.
As an example, we’ll use one of the datasets supplied with gretl, that is rac3d, which contains an
example taken from Cameron and Trivedi (2013) on the health conditions of 5190 people. The
script below compares the unconditional (marginal) distribution of the number of illnesses in the
past 2 weeks with the distribution of the same variable, conditional on age classes.
open rac3d.gdt
# unconditional boxplot
boxplot ILLNESS --output=display
# create a discrete variable for age class:
# 0 = below 20, 1 = between 20 and 39, etc
series age_class = floor(AGE/0.2)
# conditional boxplot
boxplot ILLNESS age_class --factorized --output=display
After running the code above, you should see two graphs similar to Figure 6.5. By comparing the
marginal plot to the factorized one, the effect of age on the mean number of illnesses is quite
Chapter 6. Graphs and plots 46
evident: by joining the green crosses you get what is technically known as the conditional mean
function, or regression function if you prefer.
5 5
4
4
ILLNESS
3
2
2
1
1
0 0 1 2 3
age_class
7.1 Introduction
Gretl provides two commands for adding data from file to an existing dataset in the program’s
workspace, namely append and join. The append command, which has been available for a long
time, is relatively simple and is described in the Gretl Command Reference. Here we focus on the
join command, which is much more flexible and sophisticated. This chapter gives an overview
of the functionality of join along with a detailed account of its syntax and options. We provide
several toy examples and discuss one real-world case at length.
First, a note on terminology: in the following we use the terms “left-hand” and “inner” to refer
to the dataset that is already in memory, and the terms “right-hand” and “outer” to refer to the
dataset in the file from which additional data are to be drawn.
Two main features of join are worth emphasizing at the outset:
• “Key” variables can be used to match specific observations (rows) in the inner and outer
datasets, and this match need not be 1 to 1.
• A row filter may be applied to screen out unwanted observations in the outer dataset.
As will be explained below, these features support rather complex concatenation and manipulation
of data from different sources.
A further aspect of join should be noted — one that makes this command particularly useful when
dealing with very large data files. That is, when gretl executes a join operation it does not, in gen-
eral, read into memory the entire content of the right-hand side dataset. Only those columns that
are actually needed for the operation are read in full. This makes join faster and less demanding
of computer memory than the methods available in most other software. On the other hand, gretl’s
asymmetrical treatment of the “inner” and “outer” datasets in join may require some getting used
to, for users of other packages.
47
Chapter 7. Joining data sources 48
on the left, any observations outside of the current sample are set to NA; if it exists already then
observations outside of the current sample are left unchanged.
The case where you want to rename a series on import is handled by the --data option. This option
has one required argument, the name by which the series is known on the right. At this point we
need to explain something about right-hand variable names (column headings).
Right-hand names
We accept on input arbitrary column heading strings, but if these strings do not qualify as valid
gretl identifiers they are automatically converted, and in the context of join you must use the
converted names. A gretl identifier must start with a letter, contain nothing but (ASCII) letters,
digits and the underscore character, and must not exceed 31 characters. The rules used in name
conversion are:
2. Until the 31-character is reached or the input is exhausted: transcribe “legal” characters; skip
“illegal” characters apart from spaces; and replace one or more consecutive spaces with an
underscore, unless the last character transcribed is an underscore in which case space is
skipped.
In the unlikely event that this policy yields an empty string, we replace the original with coln,
where n is replaced by the 1-based index of the column in question among those used in the
join operation. If you are in doubt regarding the converted name of a given column, the function
fixname() can be used as a check: it takes the original string as an argument and returns the
converted name. Examples:
? eval fixname("valid_identifier")
valid_identifier
? eval fixname("12. Some name")
Some_name
Returning to the use of the --data option, suppose we have a column headed "12. Some name"
on the right and wish to import it as x. After figuring how the right-hand name converts, we can do
No right-hand names?
Some data files have no column headings; they jump straight into the data (and you need to deter-
mine from accompanying documentation what the columns represent). Since gretl expects column
headings, you have to take steps to get the importation right. It is generally a good idea to insert a
suitable header row into the data file. However, if for some reason that’s not practical, you should
give the --no-header option, in which case gretl will name the columns on the right as col1, col2
and so on. If you do not do either of these things you will likely lose the first row of data, since
gretl will attempt to make variable names out of it, as described above.
7.3 Filtering
Rows from the outer dataset can be filtered using the --filter option. The required parameter
for this option is a Boolean condition, that is, an expression which evaluates to non-zero (true,
include the row) or zero (false, skip the row) for each of the outer rows. The filter expression may
include any of the following terms: up to three “right-hand” series (under their converted names as
Chapter 7. Joining data sources 49
explained above); scalar or string variables defined “on the left”; any of the operators and functions
available in gretl (including user-defined functions); and numeric or string constants.
Here are a few simple examples of potentially valid filter options (assuming that the specified right-
hand side columns are found):
# 5. compound condition
--filter="x < 100 && (x > 0 || y > 0)"
Note that if you are comparing against a string constant (as in example 3 above) it is necessary
to put the string in “escaped” double-quotes (each double-quote preceded by a backslash) so the
interpreter knows that F is not supposed to be the name of a variable.
It is safest to enclose the whole filter expression in double quotes, however this is not strictly
required unless the expression contains spaces or the equals sign.
In general, an error is flagged if a missing value is encountered in a series referenced in a filter
expression. This is because the condition then becomes indeterminate; taking example 2 above, if
the nkids value is NA on any given row we are not in a position to evaluate the condition nkids>2.
However, you can use the missing() function — or ok(), which is a shorthand for !missing()— if
you need a filter that keys off the missing or non-missing status of a variable.
The basic key option is named ikey; this indicates “inner key”, that is, the key variable found in the
left-hand or inner dataset. By default it is assumed that the right-hand dataset contains a column of
the same name, though as we’ll see below that assumption can be overridden. The join command
above says, find a series named xh in the right-hand dataset and add it to the left-hand one, using
the values of hid to match rows. Looking at the data in Table 7.1 we can see how this should
work. Persons 1 and 2 are both members of household 1, so they should both get values of 1 for
xh; persons 3 and 4 are members of household 2, so that xh = 4; and so on. Note that the order
Chapter 7. Joining data sources 50
in which the key values occur on the right-hand side does not matter. The gretl output from the
print command is shown in the lower panel of Table 7.1.
people.csv hholds.csv
pid,hid,gender,age,xp hid,country,xh
1,1,M,50,1 1,US,1
2,1,F,40,2 6,IT,12
3,2,M,30,3 3,UK,6
4,2,F,25,2 4,IT,8
5,3,M,40,3 2,US,4
6,4,F,35,4 5,IT,10
7,4,M,70,3
8,4,F,60,3
9,5,F,20,4
10,6,M,40,4
pid hid xh
1 1 1
2 1 1
3 2 4
4 2 4
5 3 6
6 4 8
7 4 8
8 4 8
9 5 10
10 6 12
Table 7.1: Two linked CSV data files, and the effect of a join
Note that key variables are treated conceptually as integers. If a specified key contains fractional
values these are truncated.
Two extensions of the basic key mechanism are available.
• If the outer dataset contains a relevant key variable but it goes under a different name from
the inner key, you can use the --okey option to specify the outer key. (As with other right-
hand names, this does not have to be a valid gretl identifier.) So, for example, if hholds.csv
contained the hid information, but under the name HHOLD, the join command above could
be modified as
join hholds.csv xh --ikey=hid --okey=HHOLD
• If a single key is not sufficient to generate the matches you want, you can specify a double key
in the form of two series names separated by a comma; in this case the importation of data is
restricted to those rows on which both keys match. The syntax here is, for example
Again, the --okey option may be used if the corresponding right-hand columns are named
differently. The same number of keys must be given on the left and the right, but when a
Chapter 7. Joining data sources 51
double key is used and only one of the key names differs on the right, the name that is in
common may be omitted (although the comma separator must be retained). For example, the
second of the following lines is acceptable shorthand for the first:
join foo.csv x --ikey=key1,Lkey2 --okey=key1,Rkey2
join foo.csv x --ikey=key1,Lkey2 --okey=,Rkey2
As a new variable, tmpvar will get NA for all unmatched rows; we then transcribe its values into
x. In a more complicated case one might use the smpl command to limit the sample range before
assigning tmpvar to x, or use the conditional assignment operator ?:.
One further point should be mentioned here. Given some missing values in an imported series you
may sometimes want to know whether (a) the NAs were explicitly represented in the outer data file
or (b) they arose due to “no match”. You can find this out by using a method described in the
following section, namely the count variant of the aggregation option: this will give you a series
with 0 values for all and only unmatched rows.
7.5 Aggregation
In the case of 1 to n matching of rows (n > 1) the user must specify an “aggregation method”; that
is, a method for mapping from n rows down to one. This is handled by the --aggr option which
requires a single argument from the following list:
Chapter 7. Joining data sources 52
Note that the count aggregation method is special, in that there is no need for a “data series” on
the right; the imported series is simply a function of the specified key(s). All the other methods
require that “actual data” are found on the right. Also note that when count is used, the value
returned when no match is found is (as one might expect) zero rather than NA.
The basic use of the seq method is shown above: following the colon you give a positive integer rep-
resenting the (1-based) position of the observation in the sequence of matched rows. Alternatively,
a negative integer can be used to count down from the last match (seq:-1 selects the last match,
seq:-2 the second-last match, and so on). If the specified sequence number is out of bounds for a
given observation this method returns NA.
Referring again to the data in Table 7.1, suppose we want to import data from the persons file into
a dataset established at household level. Here’s an example where we use the individual age data
from people.csv to add the average and minimum age of household members.
Here’s a further example where we add to the household data the sum of the personal data xp, with
the twist that we apply filters to get the sum specifically for household members under the age of
40, and for women.
The possibility of using an auxiliary variable with the min and max modes of aggregation gives extra
flexibility. For example, suppose we want for each household the income of its oldest member:
country,GDP
UK,100
US,500
IT,150
FR,180
The variable country, which is also found in hholds.csv, is string-valued. We can pull the GDP of
the country in which the household resides into our households dataset with
open hholds.csv -q
join countries.csv GDP --ikey=country
which gives
1 1 1 500
2 6 2 150
3 3 3 100
4 4 2 150
5 2 1 500
6 5 2 150
1. The varname field in the command can take the form of a space-separated list of names rather
than a single name.
2. Alternatively, you can give the name of an array of strings in place of varname: the elements
of this array should be the names of the series to import.
1. The --data option, which permits the renaming of a series on import, is not available. When
importing multiple series you are obliged to accept their “outer” names, fixed up as described
in section 7.2.
2. While the other join options are available, they necessarily apply uniformly to all the series
imported via a given command. This means that if you want to import several series but using
different keys, filters or aggregation methods you must use a sequence of commands.
We then restrict the sample to married women from 30 to 60 years of age, and additionally restrict
the sample of women to those who are either heads of households or spouses of the head.
For compatibility with the Mroz dataset as presented in the gretl data file mroz87.gdt, we rename
the age and education variables as WA and WE respectively, we compute the CIT dummy and finally
we store the reduced base dataset in gretl format.
rename ETA WA
rename STUDIO WE
series CIT = (ACOM4C > 2)
store mroz_rep.gdt
The next step will be to get data on working hours from the jobs file allb1.csv. There’s a com-
plication here. We need the total hours worked over the course of the year (for both the women
and their husbands). This is not available as such, but the variables ORETOT and MESILAV give,
respectively, average hours worked per week and the number of months worked in 2010, each on a
per-job basis. If each person held at most one job over the year we could compute his or her annual
hours as
The ASCII (CSV) data files for the 2010 survey are available at http://www.bancaditalia.it/statistiche/indcamp/
bilfait/dismicro/annuale/ascii/ind10_ascii.zip.
Chapter 7. Joining data sources 55
However, some people had more than one job, and in this case what we want is the sum of annual
hours across their jobs. We could use join with the seq aggregation method to construct this sum,
but it is probably more straightforward to read the allb1 data, compute the HRS values per job as
shown above, and save the results to a temporary CSV file.
Now we can reopen the base dataset and join the hours variable from HRS.csv. Note that we need
a double key here: the women are uniquely identified by the combination of NQUEST and NORD. We
don’t need an okey specification since these keys go under the same names in the right-hand file.
We define labor force participation, LFP, based on hours.
open mroz_rep.gdt
join HRS.csv WHRS --ikey=NQUEST,NORD --data=HRS --aggr=sum
WHRS = misszero(WHRS)
LFP = WHRS > 0
For reference, here’s how we could have used seq to avoid writing a temporary file:
To generate the work experience variable, AX, we use the file lavoro.csv: this contains a variable
named ETALAV which records the age at which the person first started work.
We compute the woman’s hourly wage, WW, as the ratio of total employment income to annual
working hours. This requires drawing the series YL (payroll income) and YM (net self-employment
income) from the persons file rper10.csv.
The family’s net disposable income is available as Y in the file rfam10.csv; we import this as
FAMINC.
Data on number of children are now obtained by applying the count method. For the Mroz repli-
cation we want the number of children under the age of 6, and also the number aged 6 to 18.
We want to add data on the women’s husbands, but how do we find them? To do this we create an
additional inner key which we’ll call H_ID (husband ID), by sub-sampling in turn on the observations
Chapter 7. Joining data sources 56
falling into each of two classes: (a) those where the woman is recorded as head of household and
(b) those where the husband has that status. In each case we want the individual ID (NORD) of the
household member whose status is complementary to that of the woman in question. So for case
(a) we subsample using PARENT==1 (head of household) and filter the join using PARENT==2 (spouse
of head); in case (b) we do the converse. We thus construct H_ID piece-wise.
Now we can use our new inner key to retrieve the husbands’ data, matching H_ID on the left with
NORD on the right within each household.
The remainder of the script is straightforward and does not require discussion here: we recode
the education variables for compatibility; delete some intermediate series that are not needed any
more; add informative labels; and save the final product. See the Appendix for details.
To compare the results from this dataset with those from the earlier US data used by Mroz, one can
copy the input file heckit.inp (supplied with the gretl package) and substitute mroz_rep.gdt for
mroz87.gdt. It turns out that the results are qualitatively very similar.
Some additional points regarding dates will be taken up as they become relevant in practical cases
of joining data.
• The default assumptions are: (1) the time key appears in the first column; (2) the heading
of this column is either left blank or is one of obs, date, year, period, observation, or
observation_date (on a case-insensitive comparison); and (3) the time format conforms
to ISO 8601 where applicable (“extended” daily date format YYYY-MM-DD, monthly format
YYYY-MM, or annual format YYYY).
• If dates do not appear in the first column of the outer file, or if the column heading or format
is not as just described, the --tkey option can be used to indicate which column should be
used and/or what format should be assumed.
--tkey="Period,%m/%d/%Y"
--tkey="Period"
--tkey="obsperiod"
--tkey=",%Ym%m"
The first of these applies if Period is not the first column on the right, and dates are given in the
US format of month, day, year, separated by slashes. The second implies that although Period is
not the first column, the date format is ISO 8601. The third again implies that the date format is
OK; here the name is required even if obsperiod is the first column since this heading is not one
recognized by gretl’s heuristic. The last example implies that dates are in the first column (with
one of the recognized headings), but are given in the non-standard format year, “m”, month.
The date format string should be composed using the codes employed by the POSIX function
strptime; Table 7.2 contains a list of the most relevant codes.3
3 The %q code for quarter is not present in strptime; it is added for use with join since quarterly data are common
in macroeconomics.
Chapter 7. Joining data sources 58
Code Meaning
%% The % character.
%b The month name according to the current locale, either abbreviated
or in full.
%C The century number (0–99).
%d The day of month (1–31).
%D Equivalent to %m/%d/%y. (This is the American style date, very con-
fusing to non-Americans, especially since %d/%m/%y is widely used in
Europe. The ISO 8601 standard format is %Y-%m-%d.)
%H The hour (0–23).
%j The day number in the year (1–366).
%m The month number (1–12).
%n Arbitrary whitespace.
%q The quarter (1–4).
%w The weekday number (0–6) with Sunday = 0.
%y The year within century (0–99). When a century is not otherwise spec-
ified, values in the range 69–99 refer to years in the twentieth century
(1969–1999); values in the range 00–68 refer to years in the twenty-
first century (2000–2068).
%Y The year, including century (for example, 1991).
Date,Open,High,Low,Close,Volume,Adj Close
2013-08-02,195.50,195.50,193.22,195.16,3861000,195.16
2013-08-01,196.65,197.17,195.41,195.81,2856900,195.81
2013-07-31,194.49,196.91,194.49,195.04,3810000,195.04
Note that the data are in reverse time-series order — that won’t matter to join, the data can appear
in any order. Also note that the first column is headed Date and holds daily dates as ISO 8601
extended. That means we can pull the data into gretl very easily. In the following fragment we
create a suitably dimensioned empty daily dataset then rely on the default behavior of join with
time-series data to import the closing stock price.
nulldata 500
setobs 5 2012-01-01
join IBM.csv Close
To make explicit what we’re doing, we could accomplish exactly the same using the --tkey option:
to the present example and can be omitted without altering the sample script.
Chapter 7. Joining data sources 59
Frequency,Period,Country,Value,Flags
"Quarterly","Q1-1960","France",463876.148126845,E
"Quarterly","Q1-1960","Germany",768802.119278467,E
"Quarterly","Q1-1960","Italy",414629.791450547,E
"Quarterly","Q1-1960","United Kingdom",578437.090291889,E
"Quarterly","Q2-1960","France",465618.977328614,E
"Quarterly","Q2-1960","Germany",782484.138122549,E
"Quarterly","Q2-1960","Italy",420714.910290157,E
"Quarterly","Q2-1960","United Kingdom",572853.474696578,E
"Quarterly","Q3-1960","France",469104.41925852,E
"Quarterly","Q3-1960","Germany",809532.161494483,E
"Quarterly","Q3-1960","Italy",426893.675840156,E
"Quarterly","Q3-1960","United Kingdom",581252.066618986,E
"Quarterly","Q4-1960","France",474664.327992619,E
"Quarterly","Q4-1960","Germany",817806.132384948,E
"Quarterly","Q4-1960","Italy",427221.338414114,E
...
Table 7.3: Example of CSV file as provided by the OECD statistical website
This is an instance of data in what we call atomic format, that is, a format in which each line of the
outer file contains a single data-point and extracting data mainly requires filtering the appropriate
lines. The outer time key is under the Period heading, and has the format Q<quarter>-<year>.
Assuming that the file in Table 7.3 has the name oecd.csv, the following script reconstructs the
time series of Gross Domestic Product for several countries:
nulldata 220
setobs 4 1960:1
Note the use of the format codes %q for the quarter and %Y for the 4-digit year. A touch of elegance
could have been added by storing the invariant options to join using the setopt command, as in
If one were importing a large number of such series it might be worth rewriting the sequence of
joins as a loop, as in
• If a specified column is not selected for a substantive role in the join operation (as data to be
imported, as a key, or as an auxiliary variable for use in aggregation) the column in question
is not read and so no conversion is carried out.
• If a specified column contains numerical rather than string values, no conversion is carried
out.
• If a string value in a selected column fails parsing using the relevant time format (user-
specified or default), the converted value is NA.
• On successful conversion, the output is always in daily-date form as stated above. If you
specify a monthly or quarterly time format, the converted date is the first day of the month
or quarter.
Frequency,Period,Country,Value,Flags
"Quarterly","Q1-1960","France",463876.148126845,E
"Quarterly","Q1-1960","Germany",768802.119278467,E
"Quarterly","Q1-1960","Italy",414629.791450547,E
"Quarterly","Q1-1960","United Kingdom",578437.090291889,E
"Quarterly","Q2-1960","France",465618.977328614,E
Chapter 7. Joining data sources 61
and so on. If we have four countries and quarterly observations running from 1960:1 to 2013:2 (T
= 214 quarters) we might set up our panel workspace like this:
scalar N = 4
scalar T = 214
scalar NT = N*T
nulldata NT --preserve
setobs T 1.1 --stacked-time-series
The relevant outer keys are obvious: Country for the country and Period for the time period. Our
task is now to construct matching keys in the inner dataset. This can be done via two panel-specific
options to the setobs command. Let’s work on the time dimension first:
This variant of setobs allows us to tell gretl that time in our panel is quarterly, starting in the
first quarter of 1960. Having set that, the accessor $obsdate will give us a series of 8-digit dates
representing the first day of each quarter — 19600101, 19600401, 19600701, and so on, repeating
for each country. As we explained in section 7.11, we can use the --tconvert option on the outer
series Period to get exactly matching values (in this case using a format of Q%q-%Y for parsing the
Period values).
Now for the country names:
Here we write into the string cstrs the names of the countries, using escaped double-quotes to
handle the space in “United Kingdom”, then pass this string to setobs with the --panel-groups
option, preceded by the identifier country. This asks gretl to construct a string-valued series
named country, in which each name will repeat T times.
We’re now ready to join. Assuming the OECD file is named oecd.csv we do
# Real GDP
Period,France,Germany,Italy,"United Kingdom"
"Q1-1960",463863,768757,414630,578437
"Q2-1960",465605,782438,420715,572853
"Q3-1960",469091,809484,426894,581252
"Q4-1960",474651,817758,427221,584779
"Q1-1961",482285,826031,442528,594684
...
Call this file side_by_side.csv. Assuming the same initial set-up as above, we can panelize the
data by setting the sample to each country’s time series in turn and importing the relevant column.
The only point to watch here is that the string “United Kingdom”, being a column heading, will
become United_Kingdom on importing (see section 7.2) so we’ll need a slightly different set of
country strings.
Chapter 7. Joining data sources 62
If our working dataset and the outer data file are dimensioned such that there are just as many
time-series observations on the right as there are time slots on the left — and the observations
on the right are contiguous, in chronological order, and start on the same date as the working
dataset—we could dispense with the key apparatus and just use the first line of the join command
shown above. However, in general it is safer to use keys to ensure that the data end up in correct
registration.
flag effect
--data Give the name of the data column on the right, in case it differs from
varname (7.2); single import only
--filter Specify a condition for filtering data rows (7.3)
--ikey Specify up to two keys for matching data rows (7.4)
--okey Specify outer key name(s) in case they differ the inner ones (7.4)
--aggr Select an aggregation method for 1 to n joins (7.5)
--tkey Specify right-hand time key (7.10)
--tconvert Select outer date columns for conversion to numeric form (7.11)
--tconv-fmt Specify a format for use with tconvert (7.11)
--no-header Treat the first row on the right as data (7.2)
--verbose Report on progress in reading the outer data
Chapter 7. Joining data sources 63
Realtime data
8.1 Introduction
As of gretl version 1.9.13 the join command (see chapter 7) has been enhanced to deal with so-
called realtime datasets in a straightforward manner. Such datasets contain information on when
the observations in a time series were actually published by the relevant statistical agency and how
they have been revised over time. Probably the most popular sources of such data are the “Alfred”
online database at the St. Louis Fed (http://alfred.stlouisfed.org/) and the OECD’s StatEx-
tracts site, http://stats.oecd.org/. The examples in this chapter deal with files downloaded
from these sources, but should be easy to adapt to files with a slightly different format.
As already stated, join requires a column-oriented plain text file, where the columns may be sepa-
rated by commas, tabs, spaces or semicolons. Alfred and the OECD provide the option to download
realtime data in this format (tab-delimited files from Alfred, comma-delimited from the OECD). If
you have a realtime dataset in a spreadsheet file you must export it to a delimited text file before
using it with join.
Representing revision histories is more complex than just storing a standard time series, because
for each observation period you have in general more than one published value over time, along
with the information on when each of these values were valid or current. Sometimes this is repre-
sented in spreadsheets with two time axes, one for the observation period and another one for the
publication date or “vintage”. The filled cells then form an upper triangle (or a “guillotine blade”
shape, if the publication dates do not reach back far enough to complete the triangle). This format
can be useful for giving a human reader an overview of realtime data, but it is not optimal for
automatic processing; for that purpose “atomic” format is best.
1 If you choose to download in Excel format from OECD you get a file in the triangular or guillotine format mentioned
above.
2 In the Alfred file we have used commas rather than tabs as the column delimiter; in the OECD example we have
65
Chapter 8. Realtime data 66
two columns, which jointly record the interval over which a given data vintage was current. Daily
dates might, however, be considered overly precise for the first column, since the data period may
well be the year, quarter or month (as it is in fact here). However, following Alfred’s practice it is
acceptable to specify a daily date, indicating the first day of the period, even for non-daily data.3
Compare the first data line of the OECD example. There’s a greater amount of leading metadata,
which is left implicit in the Alfred file. Here Time is the equivalent of Alfred’s observation_date,
and Edition the equivalent of Alfred’s realtime_start_date. So we read that in February 1999
a value of 100 was current for the UK index of industrial production for January 1990, and from
the next line we see that in the same vintage month a value of 99.3 was current for industrial
production in February 1990.
Besides the different names and ordering of the columns, there are a few more substantive differ-
ences between Alfred and OECD files, most of which are irrelevant for join but some of which are
(possibly) relevant.
The first (irrelevant) difference is the ordering of the lines. It appears (though we’re not sure how
consistent this is) that in Alfred files the lines are sorted by observation date first and then by
publication date—so that all revisions of a given observation are grouped together —while OECD
files are sorted first by revision date (Edition) and then by observation date (Time). If we want the
next revision of UK industrial production for January 1990 in the OECD file we have to scan down
several lines until we find
This difference is basically irrelevant because join can handle the case where the lines appear in
random order, although some operations can be coded more conveniently if we’re able to assume
chronological ordering (either on the Alfred or the OECD pattern, it doesn’t matter).
The second (also irrelevant) difference is that the OECD seems to include periodic “Edition” lines
even when there is no change from the previous value (as illustrated above, where the UK industrial
production index for January 1990 is reported as 100 as of March 1999, the same value that we saw
to be current in February 1999), while Alfred reports a new value only when it differs from what
was previously current.
A third difference lies in the dating of the revisions or editions. As we have seen, Alfred gives
a specific daily date while (in the UK industrial production file at any rate), the OECD just dates
each edition to a month. This is not necessarily relevant for join, but it does raise the question of
whether the OECD might date revisions to a finer granularity in some of their files, in which case
one would have to be on the lookout for a different date format.
The final difference is that Alfred supplies an “end date” for each data vintage while the OECD
3 Notice that this implies that in the Alfred example it is not clear without further information whether the observation
period is the first quarter of 1960, the month January 1960, or the day January 1st 1960. However, we assume that this
information is always available in context.
Chapter 8. Realtime data 67
supplies only a starting date. But there is less to this difference than meets the eye: according to
the Alfred webmaster, “by design, a new vintage must start immediately following (the day after)
the lapse of the old vintage”—so the end date conveys no independent information.4
1. If a format is given with the --tkey option it always applies to the tkey column alone; and
for that column it overrides any format given via the --tconv-fmt option.
2. If a format is given via tconv-fmt it is assumed to apply to all the tconvert columns, unless
this assumption is overriden by rule 1.
nulldata 132
setobs 12 2004:01
For convenience we can put the name of our realtime file into a string variable. On Windows this
might look like
4 Email received from Travis May of the Federal Reserve Bank of St. Louis, 2013-10-17. This closes off the possibility
that a given vintage could lapse or expire some time before the next vintage becomes available, hence giving rise to a
“hole” in an Alfred realtime file.
Chapter 8. Realtime data 68
We can then import the data vintage 2011-06-15 using join, arbitrarily choosing the self-explanatory
identifier ip_asof_20110615.
• The --tkey option specifies the column which should be treated as holding the observation
period identifiers to be matched against the periods in the current gretl dataset.5 The more
general form of this option is --tkey="colname,format" (note the double quotes here), so
if the dates do not come in standard format, we can tell gretl how to parse them by using
the appropriate conversion specifiers as shown in Table 7.2. For example, here we could have
written --tkey="observation_date,%Y-%m-%d".
• Next, --data=INDPRO tells gretl that we want to retrieve the entries stored in the column
named INDPRO.
• As explained in section 7.11 the --tconvert option selects certain columns in the right-hand
data file for conversion from date strings to 8-digit numbers on the pattern YYYYMMDD. We’ll
need this for the next step, filtering, since the transformation to numerical values makes
it possible to perform basic arithmetic on dates. Note that since date strings in Alfred files
conform to gretl’s default assumption it is not necessary to use the --tconv-fmt option here.
• The --filter option specification in combination with the subsequent --aggr aggregation
treatment is the central piece of our data retrieval; notice how we use the date constant
20110615 in ISO basic form to do numerical comparisons, and how we perform the numerical
max operation on the converted column realtime_start_date. It would also have been
possible to predefine a scalar variable, as in
vintage = 20110615
and then use vintage in the join command instead. Here we tell join that we only want to
extract those publications that (1) already appeared before (and including) June 15th 2011,
and (2) were not yet obsoleted by a newer release.6
As a result, your dataset will now contain a time series named ip_asof_20110615 with the values
that a researcher would have had available on June 15th 2011. Of course, all values for the observa-
tions after June 2011 will be missing (and probably a few before that, too), because they only have
become available later on.
5 Strictly speaking, using --tkey is unnecessary in this example because we could just have relied on the default,
which is to use the first column in the source file for the periods. However, being explicit is often a good idea.
6 By implementing the second condition through the max aggregation on the realtime_start_date column alone,
without using the realtime_end_date column, we make use of the fact that Alfred files cannot have “holes” as explained
before.
Chapter 8. Realtime data 69
Since we do not need the realtime_start_date information for this retrieval, we have dropped
the --tconvert option here. Note that this formulation assumes that the source file is ordered
chronologically, otherwise using the option --aggr="seq:3", which retrieves the third value from
each sequence of matches, could have yielded a result different from the one intended. However,
this assumption holds for Alfred files and is probably rather safe in general.
The values of the variable imported as ip_3rdpub in this way were published at different dates,
so the variable is effectively a mix of different vintages. Depending on the type of variable, this
may also imply drastic jumps in the values; for example, index numbers are regularly re-based
to different base periods. This problem also carries over to inflation-adjusted economic variables,
where the base period of the price index changes over time. Mixing vintages in general also means
mixing different scales in the output, with which you would have to deal appropriately.7
8.6 Getting the values at a fixed lag after the observation period
New data releases may take place on any day of the month, and as we have seen the specific day
of each release is recorded in realtime files from Alfred. However, if you are working with, say,
monthly or quarterly data you may sometimes want to adjust the granularity of your realtime axis
to a monthly or quarterly frequency. For example, in order to analyse the data revision process for
monthly industrial production you might be interested in the extent of revisions between the data
available two and three months after each observation period.
This is a relatively complicated task and there is more than one way of accomplishing it. Either you
have to make several passes through the outer dataset or you need a sophisticated filter, written
as a hansl function. Either way you will want to make use of some of gretl’s built-in calendrical
functions.
We’ll assume that a suitably dimensioned workspace has been set up as described above. Given
that, the key ingredients of the join are a filtering function which we’ll call rel_ok (for “release is
OK”) and the join command which calls it. Here’s the function:
Note that we use --tconvert to convert both the observation date and the realtime start date (or
release date) to 8-digit numerical values. Both of these series are passed to the filter, which uses the
7 Some user-contributed functions may be available that address this issue, but it is beyond our scope here. Another
even more complicated issue in the realtime context is that of “benchmark revisions” applied by statistical agencies,
where the underlying definition or composition of a variable changes on some date, which goes beyond a mere rescaling.
However, this type of structural change is not, in principle, a feature of realtime data alone, but applies to any time-series
data.
Chapter 8. Realtime data 70
built-in function isoconv to extract year and month. We can then calculate dm, the “delta months”
since the observation date, for each release. The filter condition is that this delta should be no
greater than the specified lag, p.8
This filter condition may be satisfied by more than one release, but only the latest of those will
actually be the vintage that was current at the end of the n-th month after the observation period,
so we add the option --aggr=max(realtime_start_date). If instead you want to target the
release at the beginning of the n-th month you would have to use a slightly more complicated filter
function.
An illustration
Figure 8.1 shows four time series for the monthly index of US industrial production from October
2005 to June 2009: the value as of first publication plus the values current 3, 6 and 12 months out
from the observation date.9 From visual inspection it would seem that over much of this period
the Federal reserve was fairly consistently overestimating industrial production at first release and
shortly thereafter, relative to the figure they arrived at with a lag of a year.
The script that produced this Figure is shown in full in Listing 8.1. Note that in this script we are
using a somewhat more efficient version of the rel_ok function shown above, where we pass the
required series arguments in “pointer” form to avoid having to copy them (see chapter 13).
116
114
112
110
108
106
104
102
100
98
First publication
Plus 3 months
96 Plus 6 months
Plus 12 months
94
2006 2007 2008 2009
8 The filter is written on the assumption that the lag is expressed in months; on that understanding it could be used
with annual or quarterly data as well as monthly. The idea could be generalized to cover weekly or daily data without
much difficulty.
9 Why not a longer series? Because if we try to extend it in either direction we immediately run into the index re-basing
problem mentioned in section 8.5, with big (staggered) leaps downward in all the series.
Chapter 8. Realtime data 71
nulldata 45
setobs 12 2005:10
# plus 3 months
join @fname plus3 --data=INDPRO --tkey=observation_date \
--tconvert="observation_date,realtime_start_date" \
--filter="rel_ok(&observation_date, &realtime_start_date, 3)" \
--aggr=max(realtime_start_date)
# plus 6 months
join @fname plus6 --data=INDPRO --tkey=observation_date \
--tconvert="observation_date,realtime_start_date" \
--filter="rel_ok(&observation_date, &realtime_start_date, 6)" \
--aggr=max(realtime_start_date)
# plus 12 months
join @fname plus12 --data=INDPRO --tkey=observation_date \
--tconvert="observation_date,realtime_start_date" \
--filter="rel_ok(&observation_date, &realtime_start_date, 12)" \
--aggr=max(realtime_start_date)
it run to the end of the vintage history (in this file, March 2013). Our outer time key is the realtime
start date and we filter on observation date; we name the imported INDPRO values as ip_jan70.
Since it sometimes happens that more than one revision occurs in a given month we need to select
an aggregation method: here we choose to take the last revision in the month.
Recall from section 8.2 that Alfred records a new revision only when the data-point in question
actually changes. This means that our imported series will contain missing values for all months
when no real revision took place. However, we can apply a simple autoregressive procedure to fill
in the data: each missing value equals the prior non-missing value.
Figure 8.2 displays the revision history. Over this sample period the periodic re-basing of the index
overshadows amendments due to accrual of new information.
180
160
140
120
ip_jan70
100
80
60
40
20
1970 1975 1980 1985 1990 1995 2000 2005 2010
Figure 8.2: Vintages of the index of US industrial production for January 1970
Chapter 9
9.1 Introduction
The genr command provides a flexible means of defining new variables. It is documented in the
Gretl Command Reference. This chapter offers a more expansive discussion of some of the special
functions available via genr and some of the finer points of the command.
where the integer k is known as the window size and the wi terms are the so-called Bartlett weights,
|i|
defined as wi = 1 − k+1 . It can be shown that, for k large enough, ω̂2 (k)/T yields a consistent
estimator of the variance of X̄.
Gretl implements this estimator by means of the function lrvar(). This function takes one re-
quired argument, namely the series whose long-run variance is to be estimated, followed by two
optional arguments. The first of these can be used to supply a value for k; if it is omitted or nega-
tive, the popular choice T 1/3 is used. The second allows specification of an assumed value for the
population mean of X, which then replaces X̄ in the variance calculation. Usage is illustrated below.
73
Chapter 9. Special functions in genr 74
least the one-tailed, right-hand p-value as returned by the pvalue function) is the complementary
probability, the integral from x to the upper limit of the distribution, typically +∞.
In principle, therefore, there is no need for two distinct functions: given a CDF value p0 you could
easily find the corresponding p-value as 1 − p0 (or vice versa). In practice, with finite-precision
computer arithmetic, the two functions are not redundant. This requires a little explanation. In
gretl, as in most statistical programs, floating point numbers are represented as “doubles” —
double-precision values that typically have a storage size of eight bytes or 64 bits. Since there are
only so many bits available, only so many floating-point numbers can be represented: doubles do
not model the real line. Typically doubles can represent numbers over the range (roughly) ±1.7977×
10308 , but only to about 15 digits of precision.
Suppose you’re interested in the left tail of the χ 2 distribution with 50 degrees of freedom: you’d
like to know the CDF value for x = 0.9. Take a look at the following interactive session:
The cdf function has produced an accurate value, but the pvalue function gives an answer of 1,
from which it is not possible to retrieve the answer to the CDF question. This may seem surprising
at first, but consider: if the value of p1 above is correct, then the correct value for p2 is 1−8.94977×
10−35 . But there’s no way that value can be represented as a double: that would require over 30
digits of precision.
Of course this is an extreme example. If the x in question is not too far off into one or other tail
of the distribution, the cdf and pvalue functions will in fact produce complementary answers, as
shown below:
But the moral is that if you want to examine extreme values you should be careful in selecting the
function you need, in the knowledge that values very close to zero can be represented as doubles
while values very close to 1 cannot.
from a model); reset (Ramsey’s RESET); restrict (general linear restriction); runs (runs test for
randomness); testuhat (test for normality of residual); and vartest (test for difference of vari-
ances). In most cases both a $test and a $pvalue are stored; the exception is the KPSS test, for
which a p-value is not currently available.
An important point to notice about this mechanism is that the internal variables $test and $pvalue
are over-written each time one of the tests listed above is performed. If you want to reference these
values, you must do so at the correct point in the sequence of gretl commands.
A related point is that some of the test commands generate, by default, more than one test statistic
and p-value; in these cases only the last values are stored. To get proper control over the retrieval
of values via $test and $pvalue you should formulate the test command in such a way that the
result is unambiguous. This comment applies in particular to the adf and lmtest commands.
• By default, the adf command generates three variants of the Dickey–Fuller test: one based
on a regression including a constant, one using a constant and linear trend, and one using a
constant and a quadratic trend. When you wish to reference $test or $pvalue in connection
with this command, you can control the variant that is recorded by using one of the flags
--nc, --c, --ct or --ctt with adf.
• By default, the lmtest command (which must follow an OLS regression) performs several
diagnostic tests on the regression in question. To control what is recorded in $test and
$pvalue you should limit the test using one of the flags --logs, --autocorr, --squares or
--white.
jk
where ω(j, k) = 2π i n . Since the transformation is invertible, the vector x can be recovered from
f via the so-called inverse transform
n−1
1 X iω(j,k)
xk = e fj .
n j=0
The Fourier transform is used in many diverse situations on account of this key property: the
convolution of two vectors can be performed efficiently by multiplying the elements of their Fourier
transforms and inverting the result. If
n
X
zk = xj yk−j ,
j=1
then
F (z) = F (x) F (y).
That is, F (z)k = F (x)k F (y)k .
For computing the Fourier transform, gretl uses the external library fftw3: see Frigo and Johnson
(2005). This guarantees extreme speed and accuracy. In fact, the CPU time needed to perform
the transform is O(n log n) for any n. This is why the array of numerical techniques employed in
fftw3 is commonly known as the Fast Fourier Transform.
Chapter 9. Special functions in genr 76
Gretl provides two matrix functions1 for performing the Fourier transform and its inverse: fft and
ffti. In fact, gretl’s implementation of the Fourier transform is somewhat more specialized: the
input to the fft function is understood to be real. Conversely, ffti takes a complex argument and
delivers a real result. For example:
matrix x1 = { 1 ; 2 ; 3 }
# perform the transform
matrix f = fft(a)
# perform the inverse transform
matrix x2 = ffti(f)
yields
1 6 0 1
2
x1 = f =
−1.5 0.866 2
x2 =
3 −1.5 −0.866 3
where the first column of f holds the real part and the second holds the complex part. In general,
if the input to fft has n columns, the output has 2n columns, where the real parts are stored in
the odd columns and the complex parts in the even ones. Should it be necessary to compute the
Fourier transform on several vectors with the same number of elements, it is numerically more
efficient to group them into a matrix rather than invoking fft for each vector separately.
As an example, consider the multiplication of two polynomials:
a(x) = 1 + 0.5x
b(x) = 1 + 0.3x − 0.8x 2
c(x) = a(x) · b(x) = 1 + 0.8x − 0.65x 2 − 0.4x 3
The coefficients of the polynomial c(x) are the convolution of the coefficients of a(x) and b(x);
the following gretl code fragment illustrates how to compute the coefficients of c(x):
Maximum efficiency would have been achieved by grouping a and b into a matrix. The computa-
tional advantage is so little in this case that the exercise is a bit silly, but the following alternative
may be preferable for a large number of rows/columns:
Traditionally, the Fourier transform in econometrics has been mostly used in time-series analysis,
the periodogram being the best known example. Listing 9.1 shows how to compute the periodogram
of a time series via the fft function.
nulldata 50
# generate an AR(1) process
series e = normal()
series x = 0
x = 0.9*x(-1) + e
# compute the periodogram
scale = 2*pi*$nobs
X = { x }
F = fft(X)
S = sumr(F.^2)
S = S[2:($nobs/2)+1]/scale
omega = seq(1,($nobs/2))’ .* (2*pi/$nobs)
omega = omega ~ S
# compare the built-in command
pergm x
print omega
Chapter 10
10.1 Introduction
Gretl offers the following data types:
The “numerical values” mentioned above are all double-precision floating point numbers.
In this chapter we give a run-down of the basic characteristics of each of these types and also
explain their “life cycle” (creation, modification and destruction). The list and matrix types, whose
uses are relatively complex, are discussed at greater length in chapters 14 and 16 respectively.
10.2 Series
We begin with the series type, which is the oldest and in a sense the most basic type in gretl. When
you open a data file in the gretl GUI, what you see in the main window are the ID numbers, names
(and descriptions, if available) of the series read from the file. All the series existing at any point in
a gretl session are of the same length, although some may have missing values. The variables that
can be added via the items under the Add menu in the main window (logs, squares and so on) are
also series.
For a gretl session to contain any series, a common series length must be established. This is
usually achieved by opening a data file, or importing a series from a database, in which case the
length is set by the first import. But one can also use the nulldata command, which takes as it
single argument the desired length, a positive integer.
Each series has these basic attributes: an ID number, a name, and of course n numerical values.
A series may also have a description (which is shown in the main window and is also accessible
via the labels command), a “display name” for use in graphs, a record of the compaction method
used in reducing the variable’s frequency (for time-series data only) and flags marking the variable
as discrete and/or as a numeric encoding of a qualitative characteristic. These attributes can be
edited in the GUI by choosing Edit Attributes (either under the Variable menu or via right-click), or
by means of the setinfo command.
In the context of most commands you are able to reference series by name or by ID number as you
wish. The main exception is the definition or modification of variables via a formula; here you must
use names since ID numbers would get confused with numerical constants.
Note that series ID numbers are always consecutive, and the ID number for a given series will change
if you delete a lower-numbered series. In some contexts, where gretl is liable to get confused by
78
Chapter 10. Gretl data types 79
Discrete series
It is possible to mark variables of the series type as discrete. The meaning and uses of this facility
are explained in chapter 11.
String-valued series
It is generally expected that series in gretl will be “properly numeric” (on a ratio or at least an
ordinal scale), or the sort of numerical indicator variables (0/1 “dummies”) that are traditional
in econometrics. However, there is some support for “string-valued” series — see chapter 15 for
details.
10.3 Scalars
The scalar type is relatively simple: just a convenient named holder for a single numerical value.
Scalars have none of the additional attributes pertaining to series, do not have ID numbers, and
must be referenced by name. A common use of scalar variables is to record information made
available by gretl commands for further processing, as in scalar s2 = $sigmaˆ2 to record the
square of the standard error of the regression following an estimation command such as ols.
You can define and work with scalars in gretl without having any dataset in place.
In the gretl GUI, scalar variables can be inspected and their values edited via the “Icon view” (see
the View menu in the main window).
10.4 Matrices
Matrices in gretl work much as in other mathematical software (e.g. MATLAB, Octave). Like scalars
they have no ID numbers and must be referenced by name, and they can be used without any
dataset in place. Matrix indexing is 1-based: the top-left element of matrix A is A[1,1]. Matrices
are discussed at length in chapter 16; advanced users of gretl will want to study this chapter in
detail.
Matrices have two optional attribute beyond their numerical content: they may have column and/or
row names attached; these are displayed when the matrix is printed. See the colnames and
rownames functions for details.
In the gretl GUI, matrices can be inspected, analysed and edited via the Icon view item under the
View menu in the main window: each currently defined matrix is represented by an icon.
10.5 Lists
As with matrices, lists merit an explication of their own (see chapter 14). Briefly, named lists can
(and should!) be used to make command scripts less verbose and repetitious, and more easily
modifiable. Since lists are in fact lists of series ID numbers they can be used only when a dataset is
in place.
In the gretl GUI, named lists can be inspected and edited under the Data menu in the main window,
via the item Define or edit list.
10.6 Strings
String variables may be used for labeling, or for constructing commands. They are discussed in
chapter 14. They must be referenced by name; they can be defined in the absence of a dataset.
Chapter 10. Gretl data types 80
Such variables can be created and modified via the command-line in the gretl console or via script;
there is no means of editing them via the gretl GUI.
10.7 Bundles
A bundle is a container or wrapper for various sorts of objects — primarily scalars, matrices,
strings, arrays and bundles. (Yes, a bundle can contain other bundles). Secondarily, series and
lists can be placed in bundles but this is subject to important qualifications noted below.
A bundle takes the form of a hash table or associative array: each item placed in the bundle is
associated with a key which can used to retrieve it subsequently. We begin by explaining the
mechanics of bundles then offer some thoughts on what they are good for.
To use a bundle you first either “declare” it, as in
bundle foo
These two formulations are basically equivalent, in that they both create an empty bundle. The
difference is that the second variant may be reused — if a bundle named foo already exists the
effect is to empty it—while the first may only be used once in a given gretl session; it is an error to
attempt to declare a variable that already exists.
To add an object to a bundle you assign to a compound left-hand value: the name of the bundle
followed by the key. Two forms of syntax are acceptable in this context. The recommended syntax
(for most uses) is bundlename.key ; that is, the name of the bundle followed by a dot, then the key.
Both the bundle name and the key must be valid gretl identifiers.1 For example, the statement
foo.matrix1 = m
adds an object called m (presumably a matrix) to bundle foo under the key matrix1. If you wish to
make it explicit that m is supposed to be a matrix you can use the form
matrix foo.matrix1 = m
foo["matrix1"] = m
This syntax offers greater flexibility in that the key string does not have to be a valid identifier (for
example it can include spaces). In addition, when using the square bracket syntax it is possible to
use a string variable to define or access the key in question. For example:
To get an item out of a bundle, again use the name of the bundle followed by the key, as in
matrix bm = foo.matrix1
# or using the alternative notation
matrix bm = foo["matrix1"]
# or using a string variable
matrix bm = foo[s]
1 As a reminder: 31 characters maximum, starting with a letter and composed of just letters, numbers or underscore.
Chapter 10. Gretl data types 81
Note that the key identifying an object within a given bundle is necessarily unique. If you reuse an
existing key in a new assignment, the effect is to replace the object which was previously stored
under the given key. It is not required that the type of the replacement object is the same as that
of the original.
Also note that when you add an object to a bundle, what in fact happens is that the bundle acquires
a copy of the object. The external object retains its own identity and is unaffected if the bundled
object is replaced by another. Consider the following script fragment:
bundle foo
matrix m = I(3)
foo.mykey = m
scalar x = 20
foo.mykey = x
After the above commands are completed bundle foo does not contain a matrix under mykey, but
the original matrix m is still in good health.
To delete an object from a bundle use the delete command, with the bundle/key combination, as
in
delete foo.mykey
This destroys the object associated with mykey and removes the key from the hash table.
To determine whether a bundle contains an object associated with a given key, use the inbundle()
function. This takes two arguments: the name of the bundle and the key string. The value returned
by this function is an integer which codes for the type of the object (0 for no match, 1 for scalar, 2
for series, 3 for matrix, 4 for string, 5 for bundle and 6 for array). The function typestr() may be
used to get the string corresponding to this code. For example:
Besides adding, accessing, replacing and deleting individual items, the other operations that are
supported for bundles are union, printing and deletion. As regards union, if bundles b1 and b2 are
defined you can say
bundle b3 = b1 + b2
to create a new bundle that is the union of the two others. The algorithm is: create a new bundle
that is a copy of b1, then add any items from b2 whose keys are not already present in the new
bundle. (This means that bundle union is not commutative if the bundles have one or more key
strings in common.)
If b is a bundle and you say print b, you get a listing of the bundle’s keys along with the types of
the corresponding objects, as in
? print b
bundle b:
x (scalar)
mat (matrix)
inside (bundle)
Chapter 10. Gretl data types 82
Note that in the example above the bundle b nests a bundle named inside. If you want to see
what’s inside nested bundles (with a single command) you can append the --tree option to the
print command.
open data4-10
list X = const CATHOL INCOME
bundle b
b.y = ENROLL
b.X = X
eval b.y
eval b.X
• A series, as such, is inherently a member of a dataset, and a bundle can “survive” the replace-
ment or destruction of the dataset from which a series was added. It may then be impossible
(or, even if possible, meaningless) to extract a bundled series as a series. However it’s always
possible to retrieve the values of the series in the form of a matrix (column vector).
• In gretl commands that call for series arguments you cannot give a bundled series without
first extracting it. In the little example above the series ENROLL was added to bundle b under
the key y, but b.y is not itself a series (member of a dataset), it’s just an anonymous array
of values. It therefore cannot be given as, say, the dependent variable in a call to gretl’s ols
command.
• A gretl list is just an array of ID numbers of series in a given dataset, a “macro” if you like.
So as with series, there’s no guarantee that a bundled list can be extracted as a list (though it
can always be extracted as a row vector).
The points made above are illustrated in Listing 10.1. In “Case 1” we open a little dataset with just
14 cross-sectional observations and put a series into a bundle. We then open a time-series dataset
with 64 observations, while preserving the bundle, and extract the bundled series. This instance is
legal, since the stored series does not overflow the length of the new dataset (it gets written into
the first 14 observations), but it’s probably not meaningful. It’s up to the user to decide if such
operations make sense.
In “Case 2” a similar sequence of statements leads to an error (trapped by catch) because this time
the stored series will not fit into the new dataset. We can nonetheless grab the data as a vector.
In “Case 3” we put a list of three series into a bundle. This does not put any actual data values into
the bundle, just the ID numbers of the specified series, which happen to be 4, 5 and 6. We then
switch to a dataset that contains just 4 series, so the list cannot be extracted as such (IDs 5 and 6
are out of bounds). Once again, however, we can retrieve the ID numbers in matrix form if we want.
In some cases putting a gretl list as such into a bundle may be appropriate, but in others you are
better off adding the content of the list, in matrix form, as in
open data4-10
list X = const CATHOL INCOME
bundle b
matrix b.X = {X}
In this case we’re adding a matrix with three columns and as many rows as there are in the dataset;
we have the actual data, not just a reference to the data that might “go bad”. See chapter 16 for
more on this.
Chapter 10. Gretl data types 83
bundle my_bundle
fill_out_bundle(&my_bundle)
The bundle type can also be used to advantage as the return value from a packaged function, in
cases where a package writer wants to give the user the option of accessing various results. In
the gretl GUI, function packages that return a bundle are treated specially: the output window that
displays the printed results acquires a menu showing the bundled items (their names and types),
from which the user can save items of interest. For example, a function package that estimates a
model might return a bundle containing a vector of parameter estimates, a residual series and a
covariance matrix for the parameter estimates, among other possibilities.
As a refinement to support the use of bundles as a function return type, the setnote function can
be used to add a brief explanatory note to a bundled item — such notes will then be shown in the
GUI menu. This function takes three arguments: the name of a bundle, a key string, and the note.
For example
After this, the object under the key vcv in bundle b will be shown as “covariance matrix” in a GUI
menu.
10.8 Arrays
The gretl array type is intended for scripting use. Arrays have no GUI representation and they’re
unlikely to acquire one.2
A gretl array is, as you might expect, a container which can hold zero or more objects of a certain
type, indexed by consecutive integers starting at 1. It is one-dimensional. This type is implemented
by a quite “generic” back-end. The types of object that can be put into arrays are strings, matrices,
lists, bundles and arrays.3
Of gretl’s “primary” types, then, neither scalars nor series are supported by the array mechanism.
There would be little point in supporting arrays of scalars as such since the matrix type already
plays that role, and more flexibly. As for series, they have a special status as elements of a dataset
(which is in a sense an “array of series” already) and in addition we have the list type which already
functions as a sort of array for subsets of the series in a dataset.
2 However, it’s possible to save arrays “invisibly” in the context of a GUI session, by virtue of the fact that they can be
packed into bundles (see below), and bundles can be saved as part of a “session”.
3 It was not possible to nest arrays prior to version 2019d of gretl.
Chapter 10. Gretl data types 85
Creating an array
An array can be brought into existence in any of four ways: bare declaration, assignment from null,
or using one of the functions array() or defarray(). In each case one of the specific type-words
strings, matrices, lists, bundles or arrays must be used. Here are some examples:
The “bare declaration” form and the “= null” form have the same effect of creating an empty array,
but the second can be used in contexts where bare declaration is not allowed (and it can also be
used to destroy the content of an existing array and reduce it to size zero). The array() function
expects a positive integer argument and can be used to create an array of pre-given size; in this
case the elements are initialized appropriately as empty strings, null matrices, empty lists, empty
bundles or empty arrays. The defarray() function takes a variable number of arguments (one or
more), each of which may be the name of a variable of the appropriate type or an expression which
evaluates to an object of the appropriate type.
# first case
strings S = array(3)
S[2] = "string the second"
# alternative
matrices M = null
M += mnormal(T,k)
In the first method the index must (of course) be within bounds; that is, greater than zero and
no greater than the current length of the array. When the second method is used it automatically
extends the length of the array by 1.
To get hold of an element, the array index must be used:
In each case the result is an array of strings whose length is the sum of the lengths of S1 and
S2 —and similarly for the other supported types.
Chapter 10. Gretl data types 86
A couple of points are worth noting here. First, the nelem() function works to give the number of
elements in any of the “container” types (lists, arrays, bundles, matrices). Second, if you do “print
Foo” for Foo an array, you’ll see something like:
? print Foo
Array of strings, length 5
Nesting arrays
While gretl’s array structure is in itself one-dimensional you can add extra dimensions by nesting.
For example, the code below creates an array holding n arrays of m bundles.
arrays BB = array(n)
loop i=1..n -q
bundles BB[i] = array(m)
endloop
The syntax for setting or accessing any of the n × m bundles (or their members) is then on the
following pattern:
BB[i][j].m = I(3)
eval BB[i][j]
eval BB[i][j].m # or eval BB[i][j]["m"]
where the respective array subscripts are each put into square brackets.
The elements of an array of arrays must (obviously) all be arrays, but it’s not required that they
have a common content-type. For example, the following code creates an array holding an array of
matrices plus an array of strings.
arrays AA = array(2)
matrices AA[1] = array(3)
strings AA[2] = array(3)
Chapter 10. Gretl data types 87
matrices M = array(8)
# set values of M[i] here...
bundle b
b.M = M
The mutual “packability” of bundles and arrays means that it’s possible to go quite far down the
rabbit-hole. . . users are advised not to get carried away.
scalar x
series y
matrix A
and so forth. In that case the object in question is given a default initialization, as follows: a new
scalar has value NA (missing); a new series is filled with NAs; a new matrix is null (zero rows and
columns); a new string is empty; a new list has no members, new bundle and new arrays are empty.
Declaration can be supplemented by a definite initialization, as in
scalar x = pi
series y = log(x)
matrix A = zeros(10,4)
x = y/100
z = 3.5
Here the type of x will be determined automatically depending on the context. If y is a scalar,
a series or a matrix x will inherit y’s type (otherwise an error will be generated, since division is
applicable to these types only). The new variable z will “naturally” be of scalar type.
In general, however, we recommend that you state the type of a new variable explicitly. This makes
the intent clearer to a reader of the script and also guards against errors that might otherwise be
difficult to understand (i.e. a certain variable turns out to be of the wrong type for some subsequent
calculation, but you don’t notice at first because you didn’t say what type you wanted). Exceptions
to this rule might reasonably be granted for clear and simple cases where there’s little possibility
of confusion.
Modification
Typically, the values of variables of all types are modified by assignment, using the = operator with
the name of the variable on the left and a suitable value or formula on the right:
z = normal()
x = 100 * log(y) - log(y(-1))
M = qform(a, X)
Chapter 10. Gretl data types 88
By a “suitable” value we mean one that is conformable for the type in question. A gretl variable
acquires its type when it is first created and this cannot be changed via assignment; for example, if
you have a matrix A and later want a string A, you will have to delete the matrix first.
+ One point to watch out for in gretl scripting is type conflicts having to do with the names of series brought
in from a data file. For example, in setting up a command loop (see chapter 12) it is very common to call
the loop index i. Now a loop index is a scalar (typically incremented each time round the loop). If you open
a data file that happens to contain a series named i you will get a type error (“Types not conformable for
operation”) when you try to use i as a loop index.
Although the type of an existing variable cannot be changed on the fly, gretl nonetheless tries to be
as “understanding” as possible. For example if x is an existing series and you say
x = 100
gretl will give the series a constant value of 100 rather than complaining that you are trying to
assign a scalar to a series. This issue is particularly relevant for the matrix type — see chapter 16
for details.
Besides using the regular assignment operator you also have the option of using an “inflected”
equals sign, as in the C programming language. This is shorthand for the case where the new value
of the variable is a function of the old value. For example,
For scalar variables you can use a more condensed shorthand for simple increment or decrement
by 1, namely trailing ++ or -- respectively:
x = 100
x-- # x now equals 99
x++ # x now equals 100
In the case of objects holding more than one value — series, matrices and bundles — you can mod-
ify particular values within the object using an expression within square brackets to identify the
elements to access. We have discussed this above for the bundle type and chapter 16 goes into
details for matrices. As for series, there are two ways to specify particular values for modification:
you can use a simple 1-based index, or if the dataset is a time series or panel (or if it has marker
strings that identify the observations) you can use an appropriate observation string. Such strings
are displayed by gretl when you print data with the --byobs flag. Examples:
Note that with quarterly or monthly time series there is no ambiguity between a simple index
number and a date, since dates always contain a colon. With annual time-series data, however,
such ambiguity exists and it is resolved by the rule that a number in brackets is always read as a
simple index: x[1905] means the nineteen-hundred and fifth observation, not the observation for
the year 1905. You can specify a year by quotation, as in x["1905"].
Destruction
Objects of the types discussed above, with the important exception of named lists, are all destroyed
using the delete command: delete objectname.
Chapter 10. Gretl data types 89
Lists are an exception for this reason: in the context of gretl commands, a named list expands to
the ID numbers of the member series, so if you say
delete L
for L a list, the effect is to delete all the series in L; the list itself is not destroyed, but ends up
empty. To delete the list itself (without deleting the member series) you must invert the command
and use the list keyword:
list L delete
Note that the delete command cannot be used within a loop construct (see chapter 12).
Chapter 11
Discrete variables
When a variable can take only a finite, typically small, number of values, then it is said to be discrete.
In gretl, variables of the series type (only) can be marked as discrete. (When we speak of “variables”
below this should be understood as referring to series.) Some gretl commands act in a slightly
different way when applied to discrete variables; moreover, gretl provides a few commands that
only apply to discrete variables. Specifically, the dummify and xtab commands (see below) are
available only for discrete variables, while the freq (frequency distribution) command produces
different output for discrete variables.
1. From the graphical interface, select “Variable, Edit Attributes” from the menu. A dialog box
will appear and, if the variable seems suitable, you will see a tick box labeled “Treat this
variable as discrete”. This dialog box can also be invoked via the context menu (right-click on
a variable) or by pressing the F2 key.
2. From the command-line interface, via the discrete command. The command takes one or
more arguments, which can be either variables or list of variables. For example:
list xlist = x1 x2 x3
discrete z1 xlist z2
This syntax makes it possible to declare as discrete many variables at once, which cannot
presently be done via the graphical interface. The switch --reverse reverses the declaration
of a variable as discrete, or in other words marks it as continuous. For example:
discrete foo
# now foo is discrete
discrete foo --reverse
# now foo is continuous
The command-line variant is more powerful, in that you can mark a variable as discrete even if it
does not seem to be suitable for this treatment.
Note that marking a variable as discrete does not affect its content. It is the user’s responsibility
to make sure that marking a variable as discrete is a sensible thing to do. Note that if you want to
recode a continuous variable into classes, you can use gretl’s arithmetical functionality, as in the
following example:
90
Chapter 11. Discrete variables 91
nulldata 100
# generate a series with mean 2 and variance 1
series x = normal() + 2
# split into 4 classes
series z = (x>0) + (x>2) + (x>4)
# now declare z as discrete
discrete z
Once a variable is marked as discrete, this setting is remembered when you save the data file.
open greene22_2
discrete Z5 # mark Z5 as discrete
dummify Z5
The effect of the above command is to generate 5 new dummy variables, labeled DZ5_1 through
DZ5_5, which correspond to the different values in Z5. Hence, the variable DZ5_4 is 1 if Z5 equals
4 and 0 otherwise. This functionality is also available through the graphical interface by selecting
the menu item “Add, Dummies for selected discrete variables”.
The dummify command can also be used with the following syntax:
This not only creates the dummy variables, but also a named list (see section 14.1) that can be used
afterwards. The following example computes summary statistics for the variable Y for each value
of Z5:
open greene22_2
discrete Z5 # mark Z5 as discrete
list foo = dummify(Z5)
loop foreach i foo
smpl $i --restrict --replace
summary Y
endloop
smpl --full
Since dummify generates a list, it can be used directly in commands that call for a list as input, such
as ols. For example:
open greene22_2
discrete Z5 # mark Z5 as discrete
ols Y 0 dummify(Z5)
For discrete variables, frequencies are counted for each distinct value that the variable takes. For
continuous variables, values are grouped into “bins” and then the frequencies are counted for each
bin. The number of bins, by default, is computed as a function of the number of valid observations
in the currently selected sample via the rule shown in Table 11.1. However, when the command is
invoked through the menu item “Variable, Frequency Plot”, this default can be overridden by the
user.
Observations Bins
8 ≤ n < 16 5
16 ≤ n < 50 7
√
50 ≤ n ≤ 850 d ne
n > 850 29
open greene19_1
freq TUCE
discrete TUCE # mark TUCE as discrete
freq TUCE
yields
Listing 5 variables:
0) const 1) GPA 2) TUCE 3) PSI 4) GRADE
? freq TUCE
12 1 3.12% 3.12% *
14 1 3.12% 6.25% *
17 3 9.38% 15.62% ***
Chapter 11. Discrete variables 93
As can be seen from the sample output, a Doornik–Hansen test for normality is computed auto-
matically. This test is suppressed for discrete variables where the number of distinct values is less
than 10.
This command accepts two options: --quiet, to avoid generation of the histogram when invoked
from the command line and --gamma, for replacing the normality test with Locke’s nonparametric
test, whose null hypothesis is that the data follow a Gamma distribution.
If the distinct values of a discrete variable need to be saved, the values() matrix construct can be
used (see chapter 16).
where ylist and xlist are lists of discrete variables. This produces cross-tabulations (two-way
frequencies) of each of the variables in ylist (by row) against each of the variables in xlist (by
column). Or second,
xtab xlist
In the second case a full set of cross-tabulations is generated; that is, each variable in xlist is tabu-
lated against each other variable in the list. In the graphical interface, this command is represented
by the “Cross Tabulation” item under the View menu, which is active if at least two variables are
selected.
Here is an example of use:
open greene22_2
discrete Z* # mark Z1-Z8 as discrete
xtab Z1 Z4 ; Z5 Z6
which produces
[ 0] 20 91 75 93 36 315
[ 1] 28 73 54 97 34 286
[ 0] 4 36 106 70 52 45 2 315
[ 1] 3 8 48 45 37 67 78 286
[ 0] 17 60 35 45 14 171
[ 1] 31 104 94 145 56 430
[ 0] 1 8 39 47 30 32 14 171
[ 1] 6 36 115 68 59 80 66 430
Pearson’s χ 2 test for independence is automatically displayed, provided that all cells have expected
frequencies under independence greater than 10−7 . However, a common rule of thumb states that
this statistic is valid only if the expected frequency is 5 or greater for at least 80 percent of the
cells. If this condition is not met a warning is printed.
Additionally, the --row or --column options can be given: in this case, the output displays row or
column percentages, respectively.
If you want to cut and paste the output of xtab to some other program, e.g. a spreadsheet, you
may want to use the --zeros option; this option causes cells with zero frequency to display the
number 0 instead of being empty.
Chapter 12
Loop constructs
12.1 Introduction
The command loop opens a special mode in which gretl accepts a block of commands to be re-
peated zero or more times. This feature may be useful for, among other things, Monte Carlo
simulations, bootstrapping of test statistics and iterative estimation procedures. The general form
of a loop is:
By default, the genr command operates quietly in the context of a loop (without printing informa-
tion on the variable generated). To force the printing of feedback from genr you may specify the
--verbose option to loop. The --quiet option suppresses the usual printout of the number of
iterations performed, which may be desirable when loops are nested.
The --progressive option to loop modifies the behavior of the commands print and store,
and certain estimation commands, in a manner that may be useful with Monte Carlo analyses (see
Section 12.3).
The following sections explain the various forms of the loop control expression and provide some
examples of use of loops.
+ If you are carrying out a substantial Monte Carlo analysis with many thousands of repetitions, memory
capacity and processing time may be an issue. To minimize the use of computer resources, run your script
using the command-line program, gretlcli, with output redirected to a file.
95
Chapter 12. Loop constructs 96
integer; if the value is not integral, it is converted to an integer by truncation. Note that replics is
evaluated only once, when the loop is initially compiled.
While loop
A second sort of control expression takes the form of the keyword while followed by a Boolean
expression. For example,
Execution of the commands within the loop will continue so long as (a) the specified condition
evaluates as true and (b) the number of iterations does not exceed the value of the internal variable
loop_maxiter. By default this equals 100000, but you can specify a different value (or remove the
limit) via the set command (see the Gretl Command Reference).
Index loop
A third form of loop control uses an index variable, for example i.1 In this case you specify starting
and ending values for the index, which is incremented by one each time round the loop. The syntax
looks like this: loop i=1..20.
The index variable may be a pre-existing scalar; if this is not the case, the variable is created
automatically and is destroyed on exit from the loop.
The index may be used within the loop body in either of two ways: you can access the integer value
of i or you can use its string representation, $i.
The starting and ending values for the index can be given in numerical form, by reference to pre-
defined scalar variables, or as expressions that evaluate to scalars. In the latter two cases the
variables are evaluated once, at the start of the loop. In addition, with time series data you can give
the starting and ending values in the form of dates, as in loop i=1950:1..1999:4.
This form of loop control is intended to be quick and easy, and as such it is subject to certain
limitations. In particular, the index variable is always incremented by one at each iteration. If, for
example, you have
loop i=m..n
where m and n are scalar variables with values m > n at the time of execution, the index will not be
decremented; rather, the loop will simply be bypassed.
If you need more complex loop control, see the “for” form below.
The index loop is particularly useful in conjunction with the values() matrix function when some
operation must be carried out for each value of some discrete variable (see chapter 11). Consider
the following example:
open greene22_2
discrete Z8
v8 = values(Z8)
loop i=1..rows(v8)
scalar xi = v8[i]
smpl Z8==xi --restrict --replace
printf "mean(Y | Z8 = %g) = %8.5f, sd(Y | Z8 = %g) = %g\n", \
xi, mean(Y), xi, sd(Y)
endloop
1 It is common programming practice to use simple, one-character names for such variables. However, you may use
any name that is acceptable by gretl: up to 31 characters, starting with a letter, and containing nothing but letters,
numerals and the underscore character.
Chapter 12. Loop constructs 97
In this case, we evaluate the conditional mean and standard deviation of the variable Y for each
value of Z8.
Foreach loop
The fourth form of loop control also uses an index variable, in this case to index a specified list
of strings. The loop is executed once for each string in the list. This can be useful for performing
repetitive operations on a list of variables. Here is an example of the syntax:
This loop will execute three times, printing out “peach”, “pear” and “plum” on the respective itera-
tions. The numerical value of the index starts at 1 and is incremented by 1 at each iteration.
If you wish to loop across a list of variables that are contiguous in the dataset, you can give the
names of the first and last variables in the list, separated by “..”, rather than having to type all
the names. For example, say we have 50 variables AK, AL, . . . , WY, containing income levels for the
states of the US. To run a regression of income on time for each of the states we could do:
genr time
loop foreach i AL..WY
ols $i const time
endloop
This loop variant can also be used for looping across the elements in a named list (see chapter 14).
For example:
list ylist = y1 y2 y3
loop foreach i ylist
ols $i const x1 x2
endloop
Note that if you use this idiom inside a function (see chapter 13), looping across a list that has been
supplied to the function as an argument, it is necessary to use the syntax listname.$i to reference
the list-member variables. In the context of the example above, this would mean replacing the third
line with
For loop
The final form of loop control emulates the for statement in the C programming language. The
sytax is loop for, followed by three component expressions, separated by semicolons and sur-
rounded by parentheses. The three components are as follows:
1. Initialization: This is evaluated only once, at the start of the loop. Common example: setting
a scalar control variable to some starting value.
2. Continuation condition: this is evaluated at the top of each iteration (including the first). If
the expression evaluates as true (non-zero), iteration continues, otherwise it stops. Common
example: an inequality expressing a bound on a control variable.
3. Modifier: an expression which modifies the value of some variable. This is evaluated prior
to checking the continuation condition, on each iteration after the first. Common example: a
control variable is incremented or decremented.
Chapter 12. Loop constructs 98
In this example the variable r will take on the values 0.01, 0.02, . . . , 0.99 across the 99 iterations.
Note that due to the finite precision of floating point arithmetic on computers it may be necessary
to use a continuation condition such as the above, r<.991, rather than the more “natural” r<=.99.
(Using double-precision numbers on an x86 processor, at the point where you would expect r to
equal 0.99 it may in fact have value 0.990000000000001.)
Any or all of the three expressions governing a for loop may be omitted — the minimal form is
(;;). If the continuation test is omitted it is implicitly true, so you have an infinite loop unless you
arrange for some other way out, such as a break statement.
If the initialization expression in a for loop takes the common form of setting a scalar variable to
a given value, the string representation of that scalar’s value is made available within the loop via
the accessor $varname.
nulldata 50
set seed 547
series x = 100 * uniform()
# open a "progressive" loop, to be repeated 100 times
loop 100 --progressive
series u = 10 * normal()
# construct the dependent variable
series y = 10*x + u
# run OLS regression
ols y const x
# grab the coefficient estimates and R-squared
scalar a = $coeff(const)
scalar b = $coeff(x)
scalar r2 = $rsq
# arrange for printing of stats on these
print a b r2
# and save the coefficients to file
store coeffs.gdt a b
endloop
C = α + βY γ +
as presented in Greene (2000), Example 11.3. This script is included in the gretl distribution under
the name greene11_3.inp; you can find it in gretl under the menu item “File, Script files, Practice
file, Greene...”.
The option --print-final for the ols command arranges matters so that the regression results
will not be printed each time round the loop, but the results from the regression on the last iteration
will be printed when the loop terminates.
Listing 12.3 shows how a loop can be used to estimate an ARMA model, exploiting the “outer
product of the gradient” (OPG) regression discussed by Davidson and MacKinnon (1993).
Further examples of loop usage that may be of interest can be found in chapter 18.
Chapter 12. Loop constructs 100
open greene11_3.gdt
# run initial OLS
ols C 0 Y
scalar essbak = $ess
scalar essdiff = 1
scalar beta = $coeff(Y)
scalar gamma = 1
open arma.gdt
scalar c = 0
scalar a = 0.1
scalar m = 0.1
series e = 0.0
series de_c = e
series de_a = e
series de_m = e
scalar crit = 1
# log-likelihood
scalar loglik = -0.5 * sum(e^2)
print loglik
# OPG regression
ols const sc_c sc_a sc_m --print-final --no-df-corr --vcv
# show progress
printf " constant = %.8g (gradient %#.6g)\n", c, $coeff[1]
printf " ar1 coefficient = %.8g (gradient %#.6g)\n", a, $coeff[2]
printf " ma1 coefficient = %.8g (gradient %#.6g)\n", m, $coeff[3]
crit = $T - $ess
print crit
endloop
printf "\n"
printf "constant = %.8g (se = %#.6g, t = %.4f)\n", c, se_c, c/se_c
printf "ar1 coefficient = %.8g (se = %#.6g, t = %.4f)\n", a, se_a, a/se_a
printf "ma1 coefficient = %.8g (se = %#.6g, t = %.4f)\n", m, se_m, m/se_m
Chapter 13
User-defined functions
Type Description
bool scalar variable acting as a Boolean switch
int scalar variable acting as an integer
scalar scalar variable
series data series
list named list of series
matrix matrix or vector
string string variable or string literal
bundle all-purpose container (see section 10.7)
matrices array of matrices (see section 10.8)
bundles array of bundles
strings array of strings
1 An additional parameter type is available for GUI use, namely obs; this is equivalent to int except for the way it is
102
Chapter 13. User-defined functions 103
Each element in the listing of parameters must include two terms: a type specifier, and the name
by which the parameter shall be known within the function. An example follows:
Each of the type-specifiers, with the exception of list and string, may be modified by prepending
an asterisk to the associated parameter name, as in
The meaning of this modification is explained below (see section 13.4); it is related to the use of
pointer arguments in the C programming language.
• For scalar or int parameters: minimum, maximum and/or default values; or for bool pa-
rameters, just a default value.
• For optional arguments other than scalar, int and bool, the special default value null.
• For int parameters with minimum and maximum values specified, a set of strings to associate
with the allowed numerical values (value labels).
The first three of these options may be useful in many contexts; the last two may be helpful if a
function is to be packaged for use in the gretl GUI (but probably not otherwise). We now expand on
each of the options.
• The const modifier: must be given as a prefix to the basic parameter specification, as in
const matrix M
This constitutes a promise that the corresponding argument will not be modified within the
function; gretl will flag an error if the function attempts to modify the argument.
• Minimum, maximum and default values for scalar or int types: These values should di-
rectly follow the name of the parameter, enclosed in square brackets and with the individual
elements separated by colons. For example, suppose we have an integer parameter order for
which we wish to specify a minimum of 1, a maximum of 12, and a default of 4. We can write
int order[1:12:4]
If you wish to omit any of the three specifiers, leave the corresponding field empty. For
example [1::4] would specify a minimum of 1 and a default of 4 while leaving the maximum
unlimited. However, as a special case, it is acceptable to give just one value, with no colons,
in which case the value is interpreted as a default. So for example
int k[0]
designates a default value of 0 for the parameter k, with no minimum or maximum specified.
If you wished to specify a minimum of zero with no maximum or default you would have to
write
Chapter 13. User-defined functions 104
int k[0::]
For a parameter of type bool (whose values are just zero or non-zero), you can specify a
default of 1 (true) or 0 (false), as in
bool verbose[0]
• Descriptive string: This will show up as an aid to the user if the function is packaged (see
section 13.5 below) and called via gretl’s graphical interface. The string should be enclosed
in double quotes and separated from the preceding elements of the parameter specification
with a space, as in
series y "dependent variable"
• Value labels: These may be used only with int parameters for which minimum and maximum
values have been specified (so that there is a fixed number of admissible values) and the
number of labels must match the number of values. They will show up in the graphical
interface in the form of a drop-down list, making the function writer’s intent clearer when an
integer argument represents a categorical selection. A set of value labels must be enclosed in
braces, and the individual labels must be enclosed in double quotes and separated by commas
or spaces. For example:
int case[1:3:1] {"Fixed effects", "Between model", "Random effects"}
If two or more of the trailing optional fields are given in a parameter specification, they must be
given in the order shown above: min/max/default, description, value labels. Note that there is
no facility for “escaping” characters within descriptive strings or value labels; these may contain
spaces but they cannot contain the double-quote character.
Here is an example of a well-formed function specification using all the elements mentioned above:
One advantage of specifying default values for parameters, where applicable, is that in script or
command-line mode users may omit trailing arguments that have defaults. For example, myfunc
above could be invoked with just two arguments, corresponding to y and X; implicitly p = 1, c = 1
and quiet is false.
# function definition
function scalar ols_ess (series y, list xvars)
ols y 0 xvars --quiet
printf "ESS = %g\n", $ess
return $ess
end function
# main script
open data4-1
list xlist = 2 3 4
# function call (the return value is ignored here)
ols_ess(price, xlist)
The function call gives two arguments: the first is a data series specified by name and the second is
a named list of regressors. Note that while the function offers the Error Sum of Squares as a return
value, it is ignored by the caller in this instance. (As a side note here, if you want a function to
calculate some value having to do with a regression, but are not interested in the full results of the
regression, you may wish to use the --quiet flag with the estimation command as shown above.)
A second example shows how to write a function call that assigns a return value to a variable in the
caller:
# function definition
function series get_uhat (series y, list xvars)
ols y 0 xvars --quiet
return $uhat
end function
# main script
open data4-1
list xlist = 2 3 4
# function call
series resid = get_uhat(price, xlist)
Note, however, that if myfunc is already a defined function, providing a new definition automatically
overwrites the previous one, so it should rarely be necessary to delete functions explicitly.
nulldata 10
series y = normal()
series y3 = triple1(y)
print y3
triple2(&y)
print y
These two functions produce essentially the same result — the two print statements in the caller
will show the same values—but in quite different ways. The first explicitly returns a modified
version of its input (which must be a plain series): after the call to triple1, y is unaltered; it
would have been altered only if the return value were assigned back to y rather than y3. The
second function modifies its input (given as a pointer to a series) in place without actually returning
anything.
It’s worth noting that triple2 as it stands would not be considered idiomatic as a gretl function
(although it’s formally OK). The point here is just to illustrate the distinction between passing an
argument in the default way and in pointer form.
Why make this distinction? There are two main reasons for doing so: modularity and performance.
By modularity we mean the insulation of a function from the rest of the script which calls it. One of
the many benefits of this approach is that your functions are easily reusable in other contexts. To
achieve modularity, variables created within a function are local to that function, and are destroyed
when the function exits, unless they are made available as return values and these values are “picked
up” or assigned by the caller. In addition, functions do not have access to variables in “outer scope”
(that is, variables that exist in the script from which the function is called) except insofar as these
are explicitly passed to the function as arguments.
By default, when a variable is passed to a function as an argument, what the function actually “gets”
is a copy of the outer variable, which means that the value of the outer variable is not modified by
anything that goes on inside the function. This means that you can pass arguments to a function
without worrying about possible side effects; at the same time the function writer can use argument
variables as workspace without fear of disruptive effects at the level of the caller.
The use of pointers, however, allows a function and its caller to cooperate such that an outer
variable can be modified by the function. In effect, this allows a function to “return” more than one
value (although only one variable can be returned directly — see below). To indicate that a particular
object is to be passed as a pointer, the parameter in question is marked with a prefix of * in the
function definition, and the corresponding argument is marked with the complementary prefix & in
the caller. For example,
Chapter 13. User-defined functions 107
open data4-1
list xlist = 2 3 4
scalar SSR
series resid = get_uhat_and_ess(price, xlist, &SSR)
In the above, we may say that the function is given the address of the scalar variable SSR, and it
assigns a value to that variable (under the local name ess). (For anyone used to programming in C:
note that it is not necessary, or even possible, to “dereference” the variable in question within the
function using the * operator. Unadorned use of the name of the variable is sufficient to access the
variable in outer scope.)
An “address” parameter of this sort can be used as a means of offering optional information to the
caller. (That is, the corresponding argument is not strictly needed, but will be used if present). In
that case the parameter should be given a default value of null and the the function should test to
see if the caller supplied a corresponding argument or not, using the built-in function exists().
For example, here is the simple function shown above, modified to make the filling out of the ess
value optional.
If the caller does not care to get the ess value, it can use null in place of a real argument:
Alternatively, trailing function arguments that have default values may be omitted, so the following
would also be a valid call:
One limitation on the use of pointer-type arguments should be noted: you cannot supply a given
variable as a pointer argument more than once in any given function call. For example, suppose we
have a function that takes two matrix-pointer arguments,
And suppose we have two matrices, x and y, at the caller level. The call
pointfunc(&x, &y)
will generate an error. That’s because the situation inside the function would become too confusing,
with what is really the same object existing under two names.
Chapter 13. User-defined functions 108
Const parameters
Pointer-type arguments may also be useful for optimizing performance. Even if a variable is not
modified inside the function, it may be a good idea to pass it as a pointer if it occupies a lot of
memory. Otherwise, the time gretl spends transcribing the value of the variable to the local copy
may be non-negligible compared to the time the function spends doing the job it was written for.
Listing 13.1 takes this to the extreme. We define two functions which return the number of rows
of a matrix (a pretty fast operation). The first gets a matrix as argument while the second gets a
pointer to a matrix. The functions are evaluated 500 times on a matrix with 2000 rows and 2000
columns; on a typical system floating-point numbers take 8 bytes of memory, so the total size of
the matrix is roughly 32 megabytes.
Running the code in example 13.1 will produce output similar to the following (the actual numbers
of course depend on the machine you’re using):
Elapsed time:
without pointers (copy) = 2.47197 seconds,
with pointers (no copy) = 0.00378627 seconds
set stopwatch
loop 500 -q
r = rowcount1(X)
endloop
e1 = $stopwatch
set stopwatch
loop 500 -q
r = rowcount2(&X)
endloop
e2 = $stopwatch
If a pointer argument is used for this sort of purpose — and the object to which the pointer points
is not modified (is treated as read-only) by the function — one can signal this to the user by adding
the const qualifier, as shown for function rowcount2 in Listing 13.1. When a pointer argument is
qualified in this way, any attempt to modify the object within the function will generate an error.
However, combining the const flag with the pointer mechanism is technically redundant for the
Chapter 13. User-defined functions 109
following reason: if you mark a matrix argument as const then gretl will in fact pass it in pointer
mode internally (since it can’t be modified within the function there’s no downside to simply mak-
ing it available to the function rather than copying it). So in the example above we could revise the
signature of the second function as
and call it with r = rowcount2a(X), for the same speed-up relative to rowcount1.
From the caller’s point of view the second option — using the const modifier without pointer
notation—is preferable, as it allows the caller to pass an object created “on the fly”. Suppose
the caller has two matrices, A and B, in scope, and wishes to pass their vertical concatenation as an
argument. The following call would work fine:
r = rowcount2a(A|B)
To use rowcount2, on the other hand, the caller would have to create a named variable first (since
you cannot give the “address” of a anonymous object such as A|B):
matrix AB = A|B
r = rowcount2(&AB)
This requires an extra line of code, and leaves AB occupying memory after the call.
We have illustrated using a matrix parameter, but the const modifier may be used with the same
effect—namely, the argument is passed directly, without being copied, but is protected against
modification within the function—for all the types that support the pointer apparatus.
List arguments
The use of a named list as an argument to a function gives a means of supplying a function with
a set of variables whose number is unknown when the function is written — for example, sets of
regressors or instruments. Within the function, the list can be passed on to commands such as
ols.
A list argument can also be “unpacked” using a foreach loop construct, but this requires some
care. For example, suppose you have a list X and want to calculate the standard deviation of each
variable in the list. You can do:
loop foreach i X
scalar sd_$i = sd(X.$i)
endloop
Please note: a special piece of syntax is needed in this context. If we wanted to perform the above
task on a list in a regular script (not inside a function), we could do
loop foreach i X
scalar sd_$i = sd($i)
endloop
where $i gets the name of the variable at position i in the list, and sd($i) gets its standard
deviation. But inside a function, working on a list supplied as an argument, if we want to reference
an individual variable in the list we must use the syntax listname.varname. Hence in the example
above we write sd(X.$i).
This is necessary to avoid possible collisions between the name-space of the function and the name-
space of the caller script. For example, suppose we have a function that takes a list argument, and
that defines a local variable called y. Now suppose that this function is passed a list containing
Chapter 13. User-defined functions 110
a variable named y. If the two name-spaces were not separated either we’d get an error, or the
external variable y would be silently over-written by the local one. It is important, therefore, that
list-argument variables should not be “visible” by name within functions. To “get hold of” such
variables you need to use the form of identification just mentioned: the name of the list, followed
by a dot, followed by the name of the variable.
Constancy of list arguments When a named list of variables is passed to a function, the function
is actually provided with a copy of the list. The function may modify this copy (for instance, adding
or removing members), but the original list at the level of the caller is not modified.
Optional list arguments If a list argument to a function is optional, this should be indicated by
appending a default value of null, as in
In that case, if the caller gives null as the list argument (or simply omits the last argument) the
named list X inside the function will be empty. This possibility can be detected using the nelem()
function, which returns 0 for an empty list.
String arguments
String arguments can be used, for example, to provide flexibility in the naming of variables created
within a function. In the following example the function mavg returns a list containing two moving
averages constructed from an input series, with the names of the newly created variables governed
by the string argument.
open data9-9
list malist = mavg(nocars, "nocars")
print malist --byobs
The last line of the script will print two variables named nocars_2 and nocars_4. For details on
the handling of named strings, see chapter 14.
If a string argument is considered optional, it may be given a null default value, as in
we have the series known as y. It may be useful, however, to be able to determine the names of
the variables provided as arguments. This can be done using the function argname, which takes
the name of a function parameter as its single argument and returns a string. Here is a simple
illustration:
Chapter 13. User-defined functions 111
open data9-7
namefun(QNC)
Please note that this will not always work: the arguments given to functions may be anonymous
variables, created on the fly, as in somefun(log(QNC)) or somefun(CPI/100). In that case the
argname function returns an empty string. Function writers who wish to make use of this facility
should check the return from argname using the strlen() function: if this returns 0, no name was
found.
Return values
Functions can return nothing (just printing a result, perhaps), or they can return a single variable.
The return value, if any, is specified via a statement within the function body beginning with the
keyword return, followed by either the name of a variable (which must be of the type announced
on the first line of the function definition) or an expression which produces a value of the correct
type.
Having a function return a list or bundle is a way of permitting the “return” of more than one
variable. For example, you can define several series inside a function and package them as a list;
in this case they are not destroyed when the function exits. Here is a simple example, which also
illustrates the possibility of setting the descriptive labels for variables generated in a function.
open data4-1
list xlist = price sqft
list cubelist = make_cubes(xlist)
print xlist cubelist --byobs
labels
A return statement causes the function to return (exit) at the point where it appears within the
body of the function. A function may also exit when (a) the end of the function code is reached (in
the case of a function with no return value), (b) a gretl error occurs, or (c) a funcerr statement is
reached.
The funcerr keyword—which may be followed by a string enclosed in double quotes, or the name
of a string variable, or nothing—causes a function to exit with an error flagged. If a string is
provided (either literally or via a variable), this is printed on exit, otherwise a generic error message
is printed. This mechanism enables the author of a function to pre-empt an ordinary execution
error and/or offer a more specific and helpful error message. For example,
if nelem(xlist) == 0
funcerr "xlist must not be empty"
endif
Chapter 13. User-defined functions 112
However, it is recommended programming practice to have a single return point from a function
unless this is very inconvenient. The simple example above would be better written as
Error checking
When gretl first reads and “compiles” a function definition there is minimal error-checking: the
only checks are that the function name is acceptable, and, so far as the body is concerned, that you
are not trying to define a function inside a function (see Section 13.1). Otherwise, if the function
body contains invalid commands this will become apparent only when the function is called and
its commands are executed.
Debugging
The usual mechanism whereby gretl echoes commands and reports on the creation of new variables
is by default suppressed when a function is being executed. If you want more verbose output from
a particular function you can use either or both of the following commands within the function:
set echo on
set messages on
Alternatively, you can achieve this effect for all functions via the command set debug 1. Usually
when you set the value of a state variable using the set command, the effect applies only to the
current level of function execution. For instance, if you do set messages on within function f1,
which in turn calls function f2, then messages will be printed for f1 but not f2. The debug variable,
however, acts globally; all functions become verbose regardless of their level.
Further, you can do set debug 2: in addition to command echo and the printing of messages, this
is equivalent to setting max_verbose (which produces verbose output from the BFGS maximizer) at
all levels of function execution.
list xlist = 1 2 3 4
list reglist = income price
• You can use the wildcard character, “*”, to create a list of variables by name. For example,
dum* can be used to indicate all variables whose names begin with dum.
• You can use two dots to indicate a range of variables. For example income..price indicates
the set of variables whose ID numbers are greater than or equal to that of income and less
than or equal to that of price.
• If you use the keyword null on the right-hand side, you get an empty list.
• If you use the keyword dataset on the right, you get a list containing all the series in the
current dataset (except the pre-defined const).
The name of the list must start with a letter, and must be composed entirely of letters, numbers
or the underscore character. The maximum length of the name is 31 characters; list names cannot
contain spaces.
Once a named list has been created, it will be “remembered” for the duration of the gretl session
(unless you delete it), and can be used in the context of any gretl command where a list of variables
is expected. One simple example is the specification of a list of regressors:
list xlist = x1 x2 x3 x4
ols y 0 xlist
113
Chapter 14. Named lists and strings 114
Be careful: delete xlist will delete the series contained in the list, so it implies data loss (which
may not be what you want). On the other hand, list xlist delete will simply “undefine” the
xlist identifier; the series themselves will not be affected.
Similarly, to print the names of the members of a list you have to invert the usual print command,
as in
If you just say print xlist the list will be expanded and the values of all the member series will
be printed.
Lists can be modified in various ways. To redefine an existing list altogether, use the same syntax
as for creating a list. For example
list xlist = 1 2 3
xlist = 4 5 6
Another option for appending a term (or a list) to an existing list is to use +=, as in
xlist += cpi
xlist -= cpi
In most contexts where lists are used in gretl, it is expected that they do not contain any duplicated
elements. If you form a new list by simple concatenation, as in list L3 = L1 L2 (where L1 and
L2 are existing lists), it’s possible that the result may contain duplicates. To guard against this you
can form a new list as the union of two existing ones:
list L3 = L1 || L2
The result is a list that contains all the members of L1, plus any members of L2 that are not already
in L1.
In the same vein, you can construct a new list as the intersection of two existing ones:
list L3 = L1 && L2
Here L3 contains all the elements that are present in both L1 and L2.
You can also subtract one list from another:
list L3 = L1 - L2
The result contains all the elements of L1 that are not present in L2.
Chapter 14. Named lists and strings 115
matrix m = {1,2,3,4}
list L = m
Querying a list
You can determine the number of variables or elements in a list using the function nelem().
list xlist = 1 2 3
nl = nelem(xlist)
The (scalar) variable nl will be assigned a value of 3 since xlist contains 3 members.
You can determine whether a given series is a member of a specified list using the function
inlist(), as in
scalar k = inlist(L, y)
where L is a list and y a series. The series may be specified by name or ID number. The return value
is the (1-based) position of the series in the list, or zero if the series is not present in the list.
list xlist = x1 x2 x3
list lxlist = log(xlist)
list difflist = diff(xlist)
When generating a list of lags in this way, you specify the maximum lag order inside the parenthe-
ses, before the list name and separated by a comma. For example
list xlist = x1 x2 x3
list laglist = lags(2, xlist)
or
scalar order = 4
list laglist = lags(order, xlist)
These commands will populate laglist with the specified number of lags of the variables in xlist.
You can give the name of a single series in place of a list as the second argument to lags: this is
equivalent to giving a list with just one member.
The dummify function creates a set of dummy variables coding for all but one of the distinct values
taken on by the original variable, which should be discrete. (The smallest value is taken as the
omitted catgory.) Like lags, this function returns a list even if the input is a single series.
Chapter 14. Named lists and strings 116
Another useful operation you can perform with lists is creating interaction variables. Suppose you
have a discrete variable xi , taking values from 1 to n and a variable zi , which could be continuous
or discrete. In many cases, you want to “split” zi into a set of n variables via the rule
(
(j) zi when xi = j
zi =
0 otherwise;
in practice, you create dummies for the xi variable first and then you multiply them all by zi ; these
are commonly called the interactions between xi and zi . In gretl you can do
list H = D ^ Z
where D is a list of discrete series (or a single discrete series), Z is a list (or a single series)1 ; all the
interactions will be created and listed together under the name H.
An example is provided in script 14.1
series x3 = Xlist[3]
will retrieve the third element of the list Xlist under the name x3 (or will generate an error if
Xlist has less then three members).
In addition gretl offers several functions that apply to a list and return a series. In most cases,
these functions also apply to single series and behave as natural extensions when applied to lists,
but this is not always the case.
For recognizing and handling missing values, gretl offers several functions (see the Gretl Command
Reference for details). In this context, it is worth remarking that the ok() function can be used
with a list argument. For example,
list xlist = x1 x2 x3
series xok = ok(xlist)
After these commands, the series xok will have value 1 for observations where none of x1, x2, or
x3 has a missing value, and value 0 for any observations where this condition is not met.
The functions max, min, mean, sd, sum and var behave “horizontally” rather than “vertically” when
their argument is a list. For instance, the following commands
list Xlist = x1 x2 x3
series m = mean(Xlist)
produce a series m whose i-th element is the average of x1,i , x2,i and x3,i ; missing values, if any, are
implicitly discarded.
In addition, gretl provides three functions for weighted operations: wmean, wsd and wvar. Consider
as an illustration Table 14.1: the first three columns are GDP per capita for France, Germany and
Italy; columns 4 to 6 contain the population for each country. If we want to compute an aggregate
indicator of per capita GDP, all we have to do is
smpl 1 6
print D X INTER -o
1 0 2 12 32 12
2 1 1 12 30 0
3 0 2 12 35 12
4 0 1 12 34 12
5 1 2 14 31 0
6 1 0 12 54 0
1 0 32 0 0 0
2 12 0 30 0 12
3 0 35 0 0 0
4 0 34 0 0 12
5 14 0 31 0 0
6 12 0 54 12 0
1 12 0 0 32
2 0 0 30 0
3 12 0 0 35
4 0 0 34 0
5 14 0 0 31
6 0 54 0 0
so for example
Table 14.1: GDP per capita and population in 3 European countries (Source: Eurostat)
The first field after the type-name string is the name under which the string should be saved, then
comes an equals sign, then comes a specification of the string to be saved. This may take any of
the following forms:
The role of the integer offset is to use a substring of the preceding element, starting at the given
character offset. An empty string is returned if the offset is greater than the length of the string in
question.
To add to the end of an existing string you can use the operator ~=, as in
string s1 = "sweet"
string s2 = "Home, " ~ s1 ~ " home."
Note that when you define a string variable using a string literal, no characters are treated as
“special” (other than the double quotes that delimit the string). Specifically, the backslash is not
used as an escape character. So, for example,
Chapter 14. Named lists and strings 119
string s = "\"
scalar x = 8
foo = sprintf("var%d", x) # produces "var8"
• When such a variable appears among the arguments to the printf command or sprintf
function.
String substitution can be used in contexts where a string variable is not acceptable as such. If
gretl encounters the symbol @ followed directly by the name of a string variable, this notation is
treated as a “macro”: the value of the variable is sustituted literally into the command line before
the regular parsing of the command is carried out.
One common use of string substitution is when you want to construct and use the name of a series
programatically. For example, suppose you want to create 10 random normal series named norm1
to norm10. This can be accomplished as follows.
Note that plain sname could not be used in the second line within the loop: the effect would be
to attempt to overwrite the string variable named sname with a series of the same name. What
we want is for the current value of sname to be dumped directly into the command that defines a
series, and the “@” notation achieves that.
Another typical use of string substitution is when you want the options used with a particular
command to vary depending on some condition. For example,
Chapter 14. Named lists and strings 120
open data4-1
list X = const sqft
use_optstr(price, X, 1)
use_optstr(price, X, 0)
When printing the value of a string variable using the print command, the plain variable name
should generally be used, as in
But note that this next variant does something quite different.
which attempts to print the values of two variables, Just and testing.
Built-in strings
Apart from any strings that the user may define, some string variables are defined by gretl itself.
These may be useful for people writing functions that include shell commands. The built-in strings
are as shown in Table 14.2.
To access these as ordinary string variables, prepend a dollar sign (as in $dotdir); to use them in
string-substitution mode, prepend the at-sign (@dotdir).
To check whether you got a non-empty value from a given call to getenv, you can use the function
strlen, which retrieves the length of the string, as in
String-valued series
15.1 Introduction
Gretl’s support for data series with string values has gone through three phases:
1. No support: we simply rejected non-numerical values when reading data from file.
2. Numeric encoding only: we would read a string-valued series from a delimited text data file
(provided the series didn’t mix numerical values and strings) but the representation of the
data within gretl was purely numerical. We printed a “string table” showing the mapping
between the original strings and gretl’s encoding and it was up to the user to keep track of
this mapping.
3. Preservation of string values: the string table that we construct in reading a string-valued
series is now stored as a component of the dataset so it’s possible to display and manipulate
these values within gretl.
The third phase has now been in effect for several years, with a series of gradual refinements.
This chapter gives an account of the status quo. It explains how to create string-valued series and
describes the operations that are supported for such series.
city,year
"Bilbao",2009
"Toruń",2011
"Oklahoma City",2013
"Berlin",2015
"Athens",2017
122
Chapter 15. String-valued series 123
? print --byobs
city year
1 Bilbao 2009
2 Toruń 2011
3 Oklahoma C.. 2013
4 Berlin 2015
5 Athens 2017
city
1 1
2 2
3 3
4 4
5 5
• By default the print command shows us the string values of the series city, and it han-
dles non-ASCII characters provided they’re in UTF-8 (but it doesn’t handle longer strings very
elegantly).
• The --numeric option to print exposes the numeric codes for a string-valued series.
Suppose you want to access the numeric code for a particular string-valued observation: you can
get that by “casting” the series to a vector. Thus
gives
The numeric codes for string-valued series are always assigned thus: reading the data file row by
row, the first string value is assigned 1, the next distinct string value is assigned 2, and so on.
1. The series must have only integer values and the smallest value must be 1 or greater.
2. The array of strings must have at least n members, where n is the largest value found in the
series.
The logic of these conditions is that we’re looking to create a mapping as described above, from
a 1-based sequence of integers to a set of strings. However, we’re allowing for the possibility that
the series in question is an incomplete sample from an associated population. Suppose we have a
Chapter 15. String-valued series 124
series that goes 2, 3, 5, 9, 10. This is taken to be a sample from a population that has at least 10
discrete values, 1, 2, . . . , 10, and so requires at least 10 value-strings.
One aspect of stringify() is debatable. At present the function returns 0 on success, otherwise
an integer error code; it doesn’t explicitly “fail” if the required conditions are not met, and it’s up
to the user to check if things went OK. Maybe it should just fail on error?
Here’s (a simplified version of) an example that one of the authors has had cause to use: deriving
US-style “letter grades” from a series containing percentage scores for students. Call the percentage
series x, and say we want to create a series with values A for x ≥ 90, B for 80 ≤ x < 90, and so on
down to F for x < 60. Then we can do:
The way the grade series is constructed is not the most compact, but it’s nice and explicit, and
easy to amend if one wants to adjust the threshold values. Note the use of strsplit() to create
an on-the-fly array of strings from a string literal; this is convenient when the array contains a
moderate number of elements with no embedded spaces. An alternative way to get the same result
is to define the array of strings via the defarray() function, as in
stringify(grade,defarray("F","D","C","B","A"))
We should also mention that we have a function to perform the inverse operation of stringify():
the strvals() function retrieves the array of string values from a series. (It returns an empty array
if the series is not string-valued.)
grade[31] = "A"
grade[31] = 5
grade[31] += 1
What you’re not allowed to do here is make a numerical adjustment that would put the value out
of bounds in relation to the set of string values. For example, if we tried grade[31] = 6 we’d get
an error.
Chapter 15. String-valued series 125
On the other hand, you can implicitly extend the set of string values. This wouldn’t make sense for
the letter grades example but it might for, say, city names. Returning to the example in section 15.2
suppose we try
dataset addobs 1
year[6] = 2019
city[6] = "Naples?"
This will work OK: we’re implicitly adding another member to the string table for city; the associ-
ated numeric code will be the next available integer.1
Missing values
We support one exception to the general rule, never break the mapping between strings and nu-
meric codes for string-valued series: you can mark particular observations as missing. This is done
in the usual way, e.g.,
grade[31] = NA
Note, however, that on importing a string series from a delimited text file any non-blank strings (in-
cluding “NA”) will be interpreted as valid values; any missing values in such a file should therefore
be represented by blank cells.
the string values are not copied over: you get a purely numerical series holding the codes of the
original series. But if you want a full copy with the string values that can easily be arranged:
series y = sqrt(city)
1 Admittedly there is a downside to this feature: one may inadvertently add a new string value by mistyping a string
will elicit no complaint and generate a numerical series 1, 1.41421, . . . . It’s up to the user to judge
whether this sort of thing makes any sense.
Similarly, it’s up to the user to decide if it makes sense to use a string-valued series “as is” in a
regression model, whether as regressand or regressor — again, the numerical values of the series
are taken. Often this will not make sense, but sometimes it may: the numerical values may by
design form an ordinal, or even a cardinal, scale (as in the “grade” example in section 15.2).
More likely, one would want to use dummify on a string-valued series before using it in statistical
modeling. In that context gretl’s series labels are suitably informative. For example, suppose we
have a series race with numerical values 1, 2 and 3 and associated strings “White”, “Black” and
“Other”. Then the hansl code
list D = dummify(race)
labels
Given such a series you can use string values in a sample restriction, as in
if nelem(strvals(arg)) > 0
# yes
else
# no
endif
Now suppose one wanted to put something like the code that generated the grade series in sec-
tion 15.2 into a function. That can be done, but not in the form of a function that directly returns
the desired series—that is, something like
Unfortunately the above will not work: the caller will get the grade series OK but it won’t be string-
valued. At first sight this may seem to be a bug but it’s defensible as a consequence of the way
series work in gretl.
Chapter 15. String-valued series 127
The point is that series have, so to speak, two grades of existence. They can exist as fully-fledged
members of a dataset, or they can have a fleeting existence as simply anonymous arrays of numbers
that are of the same length as dataset series. Consider the statement
On the right-hand side we have the “series” x+1, which is called into existence as part of a calcula-
tion but has no name and cannot have string values. Similarly, consider
The return value from letter_grade() is likewise an anonymous array,2 incapable of holding
string values until it gets assigned to the named series grade. The solution is to define grade as a
series, at the level of the caller, before calling letter_grade(), as in
# caller
...
series grade
letter_grade(x, &grade)
As you’ll see from the account above, we don’t offer any very fancy facilities for string-valued
series. We’ll read them from suitable sources and we’ll create them natively via stringify— and
we’ll try to ensure that they retain their integrity — but we don’t, for example, take the specification
of a string-valued series as a regressor as an implicit request to include the dummification of its
distinct values. Besides laziness, this reflects the fact that in gretl a string-valued series may be
usable “as is”, depending on how it’s defined; you can use dummify if you need it.
Stata files
Stata supports two relevant sorts of variables: (1) those that are of “string type” and (2) variables
of one or other numeric type that have “value labels” defined. Neither of these is exactly equivalent
to what we call a “string-valued series” in gretl.
Stata variables of string type have no numeric representation; their values are literally strings, and
that’s all. Stata’s numeric variables with value labels do not have to be integer-valued and their
least value does not have to be 1; however, you can’t define a label for a value that is not an integer.
Thus in Stata you can have a series that comprises both integer and non-integer values, but only
the integer values can be labeled.3
This means that on import to gretl we can readily handle variables of string type from Stata’s dta
files. We give them a 1-based numeric encoding; this is arbitrary but does not conflict with any
information in the dta file. On the other hand, in general we’re not able to handle Stata’s numeric
2 A proper named series, with string values, existed while the function was executing but it ceased to exist as soon as
variables with value labels; currently we report the value labels to the user but do not attempt to
store them in the gretl dataset. We could check such variables and import them as string-valued
series if they satisfy the criteria stated in section 15.2 but we don’t at present.
Matrix manipulation
Together with the other two basic types of data (series and scalars), gretl offers a quite compre-
hensive array of matrix methods. This chapter illustrates the peculiarities of matrix syntax and
discusses briefly some of the more complex matrix functions. For a full listing of matrix functions
and a comprehensive account of their syntax, please refer to the Gretl Command Reference.
1. By direct specification of the scalar values that compose the matrix — either in numerical form,
or by reference to pre-existing scalar variables, or using computed values.
4. Via a suitable expression that references existing matrices and/or scalars, or via some special
functions.
To specify a matrix directly in terms of scalars, the syntax is, for example:
matrix A = {1, 2, 3 ; 4, 5, 6}
The matrix is defined by rows; the elements on each row are separated by commas and the rows
are separated by semi-colons. The whole expression must be wrapped in braces. Spaces within the
braces are not significant. The above expression defines a 2 × 3 matrix. Each element should be a
numerical value, the name of a scalar variable, or an expression that evaluates to a scalar. Directly
after the closing brace you can append a single quote (’) to obtain the transpose.
To specify a matrix in terms of data series the syntax is, for example,
where the names of the variables are separated by commas. Besides names of existing variables,
you can use expressions that evaluate to a series. For example, given a series x you could do
Each variable occupies a column (and there can only be one variable per column). You cannot use
the semicolon as a row separator in this case: if you want the series arranged in rows, append the
transpose symbol. The range of data values included in the matrix depends on the current setting
of the sample range.
Instead of giving an explicit list of variables, you may instead provide the name of a saved list (see
Chapter 14), as in
129
Chapter 16. Matrix manipulation 130
list xlist = x1 x2 x3
matrix A = {xlist}
When you provide a named list, the data series are by default placed in columns, as is natural in an
econometric context: if you want them in rows, append the transpose symbol.
As a special case of constructing a matrix from a list of variables, you can say
matrix A = {dataset}
This builds a matrix using all the series in the current dataset, apart from the constant (variable 0).
When this dummy list is used, it must be the sole element in the matrix definition {...}. You can,
however, create a matrix that includes the constant along with all other variables using horizontal
concatenation (see below), as in
matrix A = {const}~{dataset}
By default, when you build a matrix from series that include missing values the data rows that
contain NAs are skipped. But you can modify this behavior via the command set skip_missing
off. In that case NAs are converted to NaN (“Not a Number”). In the IEEE floating-point stan-
dard, arithmetic operations involving NaN always produce NaN. Alternatively, you can take greater
control over the observations (data rows) that are included in the matrix using the “set” variable
matrix_mask, as in
where msk is the name of a series. Subsequent commands that form matrices from series or lists will
include only observations for which msk has non-zero (and non-missing) values. You can remove
this mask via the command set matrix_mask null.
+ Names of matrices must satisfy the same requirements as names of gretl variables in general: the name
can be no longer than 31 characters, must start with a letter, and must be composed of nothing but letters,
numbers and the underscore character.
matrix A = {}
creates an empty matrix—a matrix with zero rows and zero columns.
The main purpose of the concept of an empty matrix is to enable the user to define a starting point
for subsequent concatenation operations. For instance, if X is an already defined matrix of any size,
the commands
matrix A = {}
matrix B = A ~ X
Legal operations on empty matrices are listed in Table 16.1. (All other matrix operations gener-
ate an error when an empty matrix is given as an argument.) In line with the above interpreta-
tion, some matrix functions return an empty matrix under certain conditions: the functions diag,
vec, vech, unvech when the arguments is an empty matrix; the functions I, ones, zeros,
mnormal, muniform when one or more of the arguments is 0; and the function nullspace when
its argument has full column rank.
With regard to option 2, the integer value can be given numerically, as the name of an existing
scalar variable, or as an expression that evaluates to a scalar. With option 4, the index matrix given
in the rows field must be either p × 1 or 1 × p, and should contain integer values in the range 1 to
n, where n is the number of rows in the matrix from which the selection is to be made.
The cols specification works in the same way, mutatis mutandis. Here are some examples.
matrix B = A[1,]
matrix B = A[2:3,3:5]
matrix B = A[2,2]
matrix idx = {1, 2, 6}
matrix B = A[idx,]
The first example selects row 1 from matrix A; the second selects a 2×3 submatrix; the third selects
a scalar; and the fourth selects rows 1, 2, and 6 from matrix A.
If the matrix in question is n × 1 or 1 × m, it is OK to give just one index specifier and omit the
comma. For example, A[2] selects the second element of A if A is a vector. Otherwise the comma
is mandatory.
In addition there is a pre-defined index specification, diag, which selects the principal diagonal of
a square matrix, as in B[diag], where B is square.
You can use selections of this sort on either the right-hand side of a matrix-generating formula or
the left. Here is an example of use of a selection on the right, to extract a 2 × 2 submatrix B from a
3 × 3 matrix A:
Chapter 16. Matrix manipulation 132
matrix A = {1, 2, 3; 4, 5, 6; 7, 8, 9}
matrix B = A[1:2,2:3]
And here are examples of selection on the left. The second line below writes a 2 × 2 identity matrix
into the bottom right corner of the 3 × 3 matrix A. The fourth line replaces the diagonal of A with
1s.
matrix A = {1, 2, 3; 4, 5, 6; 7, 8, 9}
matrix A[2:3,2:3] = I(2)
matrix d = {1, 1, 1}
matrix A[diag] = d
matrix A = {1, 2, 3; 4, 5, 6; 7, 8, 9}
matrix B = A[-2,-3]
which creates B as a 2 × 2 matrix which drops row 2 and column 3 from A. Negative indices can also
be given in the form of an index vector:
In this case B is formed by dropping rows 1, 3 and 5 from A (which must have at least 5 rows), but
retaining the column dimension of A.
There are two limitations on the use of negative indices. First, the from:to range syntax described
in the previous section is not available, but you can use the seq function to achieve an equivalent
effect, as in
where B drops columns 3 to 7 from A. Second, use of negative indices is valid only on the right-hand
side of a matrix calculation; there is no “negative index” equivalent of assignment to a sub-matrix,
as in
+ addition
- subtraction
* ordinary matrix multiplication
’ pre-multiplication by transpose
\ matrix “left division” (see below)
/ matrix “right division” (see below)
~ column-wise concatenation
| row-wise concatenation
** Kronecker product
== test for equality
!= test for inequality
matrix C = A + k
matrix D = A - k
both produce m × n matrices, with elements cij = aij + k and dij = aij − k respectively.
By “pre-multiplication by transpose” we mean, for example, that
matrix C = X’Y
produces the product of X-transpose and Y . In effect, the expression X’Y is shorthand for X’*Y,
which is also valid syntax. In the special case X = Y , however, the two are not exactly equivalent.
The former expression uses a specialized algorithm with two advantages: it is more efficient com-
putationally, and ensures that the result is free of machine precision artifacts that may render it
numerically non-symmetric. This, however, is unlikely to be an issue unless your X matrix is rather
large (at least several hundreds rows/columns).
In matrix “left division”, the statement
matrix X = A \ B
is interpreted as a request to find the matrix X that solves AX = B. If A is a square matrix, this is
in principle equivalent to A−1 B, which fails if A is singular; the numerical method employed here
is the LU decomposition. If A is a T × k matrix with T > k, then X is the least-squares solution,
X = (A0 A)−1 A0 B, which fails if A0 A is singular; the numerical method employed here is the QR
decomposition. Otherwise, the operation necessarily fails.
For matrix “right division”, as in X = A / B, X is the matrix that solves XB = A, in principle
equivalent to AB −1 .
In “dot” operations a binary operation is applied element by element; the result of this operation
is obvious if the matrices are of the same size. However, there are several other cases where such
operators may be applied. For example, if we write
Chapter 16. Matrix manipulation 134
matrix C = A .- B
then the result C depends on the dimensions of A and B. Let A be an m × n matrix and let B be
p × q; the result is as follows:
Case Result
Dimensions match (m = p and n = q) cij = aij − bij
A is a column vector; rows match (m = p; n = 1) cij = ai − bij
B is a column vector; rows match (m = p; q = 1) cij = aij − bi
A is a row vector; columns match (m = 1; n = q) cij = aj − bij
B is a row vector; columns match (m = p; q = 1) cij = aij − bj
A is a column vector; B is a row vector (n = 1; p = 1) cij = ai − bj
A is a row vector; B is a column vector (m = 1; q = 1) cij = aj − bi
A is a scalar (m = 1 and n = 1) cij = a − bij
B is a scalar (p = 1 and q = 1) cij = aij − b
If none of the above conditions are satisfied the result is undefined and an error is flagged.
Note that this convention makes it unnecessary, in most cases, to use diagonal matrices to perform
transformations by means of ordinary matrix multiplication: if Y = XV , where V is diagonal, it is
computationally much more convenient to obtain Y via the instruction
matrix Y = X .* v
matrix C = A ~ B
h i
produces C = A B .
Row-wise concatenation of an m × n matrix A and an p × n matrix B produces an (m + p) × n
matrix. That is,
matrix C = A | B
" #
A
produces C = .
B
Expression Effect
matrix B = A * k bij = kaij
matrix B = A / k bij = aij /k
matrix B = k / A bij = k/aij
matrix B = A + k bij = aij + k
matrix B = A - k bij = aij − k
matrix B = k - A bij = k − aij
matrix B = A % k bij = aij modulo k
matrix B = sqrt(A)
√
generates a matrix such that bij = aij . All such functions require a single matrix as argument, or
an expression which evaluates to a single matrix.1
In this section, we review some aspects of functions that apply specifically to matrices. A full
account of each function is available in the Gretl Command Reference.
Matrix reshaping
In addition to the methods discussed in sections 16.1 and 16.3, a matrix can also be created by
re-arranging the elements of a pre-existing matrix. This is accomplished via the mshape function.
It takes three arguments: the input matrix, A, and the rows and columns of the target matrix, r
and c respectively. Elements are read from A and written to the target in column-major order. If A
contains fewer elements than n = r × c, they are repeated cyclically; if A has more elements, only
the first n are used.
For example:
matrix a = mnormal(2,3)
a
matrix b = mshape(a,3,1)
b
matrix b = mshape(a,5,2)
b
produces
? a
a
? matrix b = mshape(a,3,1)
1 Note that to find the “matrix square root” you need the cholesky function (see below). And since the exp function
computes the exponential element by element, it does not return the matrix exponential unless the matrix is diagonal.
To get the matrix exponential, use mexp.
Chapter 16. Matrix manipulation 136
Generated matrix b
? b
b
1.2323
0.54363
0.99714
? matrix b = mshape(a,5,2)
Replaced matrix b
? b
b
1.2323 -0.48467
0.54363 1.2323
0.99714 0.54363
0.43928 0.99714
-0.39078 0.43928
? z1 = {1,2;3,4}
z1 = {1,2;3,4}
Generated matrix z1
? z2 = I(2)
z2 = I(2)
Generated matrix z2
? conj_z1 = z1 .* {1,-1}
conj_z1 = z1 .* {1,-1}
Generated matrix conj_z1
? eval cmult(z1,z2)
eval cmult(z1,z2)
1 2
-4 3
? eval cmult(z1,conj_z1)
eval cmult(z1,conj_z1)
5
25
The general rule is: the “main” result of the function is always returned as the result proper.
Auxiliary returns, if needed, are retrieved using pre-existing matrices, which are passed to the
function as pointers (see 13.4). If such values are not needed, the pointer may be substituted with
the keyword null.
The syntax for qrdecomp, eigensym and eigengen is of the form
The first argument, A, represents the input data, that is, the matrix whose decomposition or analysis
is required. The second argument must be either the name of an existing matrix preceded by & (to
indicate the “address” of the matrix in question), in which case an auxiliary result is written to that
matrix, or the keyword null, in which case the auxiliary result is not produced, or is discarded.
In case a non-null second argument is given, the specified matrix will be over-written with the
auxiliary result. (It is not required that the existing matrix be of the right dimensions to receive the
result.)
The function eigensym computes the eigenvalues, and optionally the right eigenvectors, of a sym-
metric n × n matrix. The eigenvalues are returned directly in a column vector of length n; if the
eigenvectors are required, they are returned in an n × n matrix. For example:
matrix V
matrix E = eigensym(M, &V)
matrix E = eigensym(M, null)
In the first case E holds the eigenvalues of M and V holds the eigenvectors. In the second, E holds
the eigenvalues but the eigenvectors are not computed.
The function eigengen computes the eigenvalues, and optionally the eigenvectors, of a general
n × n matrix. The eigenvalues are returned directly in an n × 2 matrix, the first column holding the
real components and the second column the imaginary components.
If the eigenvectors are required (that is, if the second argument to eigengen is not null), they
are returned in an n × n matrix. The column arrangement of this matrix is somewhat non-trivial:
the eigenvectors are stored in the same order as the eigenvalues, but the real eigenvectors occupy
one column, whereas complex eigenvectors take two (the real part comes first); the total number of
columns is still n, because the conjugate eigenvector is skipped. Listing 16.1 provides a (hopefully)
clarifying example (see also subsection 16.7).
The qrdecomp function computes the QR decomposition of an m × n matrix A: A = QR, where Q
is an m × n orthogonal matrix and R is an n × n upper triangular matrix. The matrix Q is returned
directly, while R can be retrieved via the second argument. Here are two examples:
matrix R
matrix Q = qrdecomp(M, &R)
matrix Q = qrdecomp(M, null)
In the first example, the triangular R is saved as R; in the second, R is discarded. The first line
above shows an example of a “simple declaration” of a matrix: R is declared to be a matrix variable
but is not given any explicit value. In this case the variable is initialized as a 1 × 1 matrix whose
single element equals zero.
The syntax for svd is
Chapter 16. Matrix manipulation 139
matrix v
A = mnormal(3,3)
/* do the eigen-analysis */
l = eigengen(A,&v)
/* eigenvalue 1 is real, 2 and 3 are complex conjugates */
print l
print v
/*
column 1 contains the first eigenvector (real)
*/
B = A*v[,1]
c = l[1,1] * v[,1]
/* B should equal c */
print B
print c
/*
columns 2:3 contain the real and imaginary parts
of eigenvector 2
*/
B = A*v[,2:3]
c = cmult(ones(3,1)*(l[2,]),v[,2:3])
/* B should equal c */
print B
print c
Chapter 16. Matrix manipulation 140
The function svd computes all or part of the singular value decomposition of the real m × n matrix
A. Let k = min(m, n). The decomposition is
A = UΣV 0
In the first case both sets of singular vectors are obtained, in the second case only the singular
values are obtained; and in the third, the right singular vectors are obtained but U is not computed.
Please note: when the third argument is non-null, it is actually V 0 that is provided. To reconstitute
the original matrix from its SVD, one can do:
This function returns the OLS estimates obtained by regressing the T × n matrix Y on the T × k
matrix X, that is, a k × n matrix holding (X 0 X)−1 X 0 Y . The Cholesky decomposition is used. The
matrix U , if not null, is used to store the residuals.
1. The first non-comment line must contain two integers, separated by a space or a tab, indicat-
ing the number of rows and columns, respectively.
Should an error occur (such as the file being badly formatted or inaccessible), an empty matrix (see
section 16.2) is returned.
2 This is not the only definition of the SVD: some writers define U as m × m, Σ as m × n (with k non-zero diagonal
elements) and V as n × n.
Chapter 16. Matrix manipulation 141
The complementary function mwrite produces text files formatted as described above. The column
separator is the tab character, so import into spreadsheets should be straightforward. Usage is
illustrated in example 16.2. Matrices stored via the mwrite command can be easily read by other
programs; the following table summarizes the appropriate commands for reading a matrix A from
a file called a.mat in some widely-used programs.3 Note that the Python example requires that the
numpy module is loaded.
nulldata 64
scalar n = 3
string f1 = "a.csv"
string f2 = "b.csv"
matrix a = mnormal(n,n)
matrix b = inv(a)
if err != 0
fprintf "Failed to write %s\n", f1
else
err = mwrite(b, f2)
endif
if err != 0
fprintf "Failed to write %s\n", f2
else
c = mread(f1)
d = mread(f2)
a = c*d
printf "The following matrix should be an identity matrix\n"
print a
endif
Optionally, the mwrite and mread functions can use gzip compression: this is invoked if the name
of the matrix file has the suffix “.gz.” In this case the elements of the matrix are written in a single
3 Matlab users may find the Octave example helpful, since the two programs are mostly compatible with one another.
Chapter 16. Matrix manipulation 142
column. Note, however, that compression should not be applied when writing matrices for reading
by third-party software unless you are sure that the software can handle compressed data.
Many of the accessors in Table 16.4 behave somewhat differently depending on the sort of model
that is referenced, as follows:
• Single-equation models: $sigma gets a scalar (the standard error of the regression); $coeff
and $stderr get column vectors; $uhat and $yhat get series.
• System estimators: $sigma gets the cross-equation residual covariance matrix; $uhat and
$yhat get matrices with one column per equation. The format of $coeff and $stderr de-
pends on the nature of the system: for VARs and VECMs (where the matrix of regressors is
the same for all equations) these return matrices with one column per equation, but for other
system estimators they return a big column vector.
• VARs and VECMs: $vcv is not available, but X 0 X −1 (where X is the common matrix of regres-
sors) is available as $xtxinv.
If the accessors are given without any prefix, they retrieve results from the last model estimated, if
any. Alternatively, they may be prefixed with the name of a saved model plus a period (.), in which
case they retrieve results from the specified model. Here are some examples:
matrix u = $uhat
matrix b = m1.$coeff
matrix v2 = m1.$vcv[1:2,1:2]
The first command grabs the residuals from the last model; the second grabs the coefficient vector
from model m1; and the third (which uses the mechanism of sub-matrix selection described above)
grabs a portion of the covariance matrix from model m1.
If the model in question a VAR or VECM (only) $compan and $vma return the companion matrix and
the VMA matrices in stacked form, respectively (see section 29.2 for details). After a vector error
correction model is estimated via Johansen’s procedure, the matrices $jalpha and $jbeta are also
Chapter 16. Matrix manipulation 143
available. These have a number of columns equal to the chosen cointegration rank; therefore, the
product
returns the reduced-rank estimate of A(1). Since β is automatically identified via the Phillips nor-
malization (see section 30.5), its unrestricted elements do have a proper covariance matrix, which
can be retrieved through the $jvbeta accessor.
scalar x = 3
matrix x = ones(2,2) # wrong!
It is possible, however, to delete or rename an existing variable then reuse the name for a variable
of a different type:
scalar x = 3
delete x
matrix x = ones(2,2) # OK
series s = x
series u1 = U[,1]
It is assumed that x and U are pre-existing matrices. In the second example the series u1 is formed
from the first column of the matrix U.
For this operation to work, the matrix (or matrix selection) must be a vector with length equal to
either the full length of the current dataset, n, or the length of the current sample range, n0 . If
n0 < n then only n0 elements are drawn from the matrix; if the matrix or selection comprises n
elements, the n0 values starting at element t1 are used, where t1 represents the starting observation
of the sample range. Any values in the series that are not assigned from the matrix are set to the
missing code.
matrix M = { listname }
Chapter 16. Matrix manipulation 144
That formulation, with the name of the list enclosed in braces, builds a matrix whose columns hold
the variables referenced in the list. What we are now describing is a different matter: if we say
matrix M = listname
(without the braces), we get a row vector whose elements are the ID numbers of the variables in the
list. This special case of matrix generation cannot be embedded in a compound expression. The
syntax must be as shown above, namely simple assignment of a list to a matrix.
To go in the other direction, you can include a matrix on the right-hand side of an expression that
defines a list, as in
list Xl = M
where M is a matrix. The matrix must be suitable for conversion; that is, it must be a row or column
vector containing non-negative integer values, none of which exceeds the highest ID number of a
series in the current dataset.
Listing 16.3 illustrates the use of this sort of conversion to “normalize” a list, moving the constant
(variable 0) to first position.
if (x[1] != 0)
scalar k = cols(x)
loop for (i=2; i<=k; i++) --quiet
if (x[i] == 0)
x[i] == x[1]
x[1] == 0
break
endif
endloop
endif
end function
open data9-7
list Xl = 2 3 0 4
matrix x = Xl
normalize_list(&x)
list Xl = x
delete M
matrix M = mnormal(100,2)
M
print M
You can get finer control on the formatting of output by using the printf command, as illustrated
in the interactive session below:
? matrix Id = I(2)
matrix Id = I(2)
Generated matrix Id
? print Id
print Id
Id (2 x 2)
1 0
0 1
? printf "%10.3f", Id
1.000 0.000
0.000 1.000
For presentation purposes you may wish to give titles to the columns of a matrix. For this you can
use the colnames function: the first argument is a matrix and the second is either a named list of
variables, whose names will be used as headings, or a string that contains as many space-separated
substrings as the matrix has columns. For example,
? matrix M = mnormal(3,3)
? colnames(M, "foo bar baz")
? print M
M (3 x 3)
open data4-1
matrix X = { const, sqft }
matrix y = { price }
matrix b = invpd(X’X) * X’y
print "estimated coefficient vector"
b
matrix u = y - X*b
scalar SSR = u’u
scalar s2 = SSR / (rows(X) - rows(b))
matrix V = s2 * inv(X’X)
V
matrix se = sqrt(diag(V))
print "estimated standard errors"
se
# compare with built-in function
ols price const sqft --vcv
Chapter 17
Calendar dates
17.1 Introduction
Any software that aims to handle time-series data must have a good built-in calendar. This is fairly
straightforward in the current era, with the Gregorian calendar now used universally for the dating
of socioeconomic observations. It is not so straightforward, however, when dealing with historical
data recorded prior to the adoption of the Gregorian calendar in place of the Julian, an event which
first occurred in the principal Catholic countries in 1582 but which took place at different dates in
different countries over a span of several centuries.
Gretl, like most data-oriented software, uses the Gregorian calendar by default for all dates, thereby
ensuring that dates are all consecutive (the latter being a requirement of the ISO 8601 standard for
dates and times).1
As readers probably know, the Julian calendar adds a leap day (February 29) on each year that is
divisible by 4 with no remainder. But this over-compensates for the fact that a 365-day year is too
short to keep the calendar synchronized with the seasons. The Gregorian calendar introduced a
more complex rule which maintains better synchronization, namely, each year divisible by 4 with
no remainder is a leap year unless it’s a centurial year (e.g. 1900) in which case it’s a leap year only
if it is divisible by 400 with no remainder. So the years 1600 and 2000 were leap years on both
calendars, but 1700, 1800, and 1900 were leap years only on the Julian calendar. While the average
length of a Julian year is 365.25 days, the Gregorian average is 365.2425 days.
The fact that the Julian calendar inserts leap days more frequently means that the Julian date
progressively (although very slowly) falls behind the Gregorian date. For example, February 18
2017 (Gregorian) is February 5 2017 on the Julian calendar. On adoption of the Gregorian calendar
it was therefore necessary to skip several days. In England, where the transition occurred in 1752,
Wednesday September 2 was directly followed by Thursday September 14.
In comparing calendars one wants to refer to a given day in terms that are not specific to either
calendar—but how to define a “given day”? This is accomplished by a count of days following some
definite event. Astronomers use the “Julian Day,” whose count starts with a particular coincidence
of astronomical cycles in the year known to the Gregorian calendar (if one extrapolates it backwards
in time) as 4714 BC. Gretl uses a similar construction as a fulcrum, but the count of what we call the
“epoch day” starts at 1 on January 1, AD 1 (that is, the first day of the Common Era), on the proleptic
Gregorian calendar.2 This is also the convention used by the GLib library, on which gretl depends
for most of its calendrical calculation. Since GLib represents epoch days as unsigned integers, this
means that gretl does not support dates prior to the Common — or, if you prefer, Christian — Era.
1 Gretl was not consistent in this regard prior to version 2017a: leap years were taken to be as defined by the Julian
calendar prior to the adoption of the Gregorian calendar by Britain and its colonies in 1752.
2 The term “proleptic,” as applied to a calendar, indicates that it is extrapolated backwards or forwards relative to its
147
Chapter 17. Calendar dates 148
A first point to note is that a daily date has three possible representations within gretl. Two
are associated with the ISO 8601 standard, namely the “extended” representation, YYYY-MM-DD,
as for example 2017-02-19, and the “basic” representation, YYYYMMDD (for example 20170219). As
mentioned above, such dates are by default taken to be relative to the (possibly proleptic) Gregorian
calendar. The third representation is as an “epoch day” (see above), for example 736379, which is
calendar-independent and can therefore be used in converting from one calendar to another.
series y, m, d
isoconv(dates, &y, &m, &d)
This is mostly just a convenience function: provided the dates input is valid on the (possibly
proleptic) Gregorian calendar it is equivalent to:
series y = floor(dates/10000)
series m = floor((dates-10000*y)/100)
series d = dates - 10000*y - 100*m
However, there is some “value added”: isoconv checks the validity of the dates input. If the
implied year, month and day for any dates observation do not correspond to a valid Gregorian
date in the Common Era,3 then all the derived series will have value NA at that observation.
The isoconv function can also handle Julian dates, should anyone have need for that facility. The
convention is that if the negative of an ISO 8601 basic date is given as an argument, the date is
taken to be on the (possibly proleptic) Julian calendar. Since dates BCE are not supported by gretl,
hopefully this should not give rise to ambiguity. The only difference from the standard usage of
isoconv is that Julian-only dates, such as 17000229, are recognized as valid. So, for example,
series y, m, d
isoconv(-dates, &y, &m, &d)
will accept a dates value of 17000229, giving (y = 1700, m = 2, d = 29), while this would give NA
values for year, month and day on the default Gregorian calendar.
edg = epochday(1700,1,1)
edj = epochday(-1700,1,1)
produces edg = 620548 and edj = 620558, indicating that the two calendars differed by 10 days at
the point in time known as January 1, 1700, on the proleptic Gregorian calendar.
Taken together with the isodate and juldate functions (which each take an epoch day argument
and return an ISO 8601 basic date on, respectively, the Gregorian and Julian calendars), epochday
3 For example, the implied month is not in the range 1–12, or the implied day is not in the range of 1 to the number
can be used to convert between the two calendars. For example, what was the date in England (still
on the Julian calendar) on the day known to Italians as June 26, 1740 (Italy having been on the
Gregorian calendar since October 1582)?
ed = epochday(1740,6,26)
english_date = juldate(ed)
printf "%.0f\n", english_date
We find that the English date was 17400615, the 15th of June. Working in the other direction, what
Italian date corresponded to the 5th of November, 1740, in England?
ed = epochday(-1740,11,5)
italian_date = isodate(ed)
printf "%.0f\n", italian_date
Answer: 17401116; Guy Fawkes night in 1740 occurred on November 16 from the Italian point of
view.
A further—and perhaps more practical — use of epoch days is checking whether daily data are
complete. Suppose we have what purport to be 7-day daily data on the Gregorian calendar with
a starting date of 2015-01-01 and an ending date of 2016-12-31. How many observations should
there be?
ed1 = epochday(2015,1,1)
ed2 = epochday(2016,12,31)
n = ed2 - ed1 + 1
We find that there should be n = 731 observations; if there are fewer, there’s something missing.
If the data are supposed to be on a 5-day week (skipping Saturday and Sunday) or 6-day week
(skipping Sunday alone) the calculation is more complicated; in this case we can use the dayspan
function, providing as arguments the epoch-day values for the first and last dates and the number
of days per week:
ed1 = epochday(2015,1,1)
ed2 = epochday(2016,12,30)
n = dayspan(ed1, ed2, 5)
Miscellaneous functions
Two additional functions that by default operate on the Gregorian calendar can be induced to
work on the Julian by the trick mentioned above, namely giving the negative of the year. These
are weekday (which takes arguments year, month and day) and monthlen (which takes arguments
month, year and days per week). Thus for example
eval weekday(-1700,2,29)
gives 4, indicating that Julian February 29, 1700 was a Thursday. And
eval monthlen(2,-1900,5)
gives 21, indicating that there were 21 weekdays in Julian February 1900.
Chapter 17. Calendar dates 150
Partial output:
ed jdate gdate hcal
Notice that although the series hcal contains the correct historical calendar (in “basic” form), the
observation labels (in the first column of the output) are still just index numbers. It may be prefer-
able to have historical dates in that role. To achieve this we can decompose the hcal series into
year, month and day, then use the special genr markers apparatus (see chapter 4). Suitable code
along with partial output is shown in Listing 17.2.
Chapter 17. Calendar dates 151
Partial output:
ed jdate gdate hcal
Partial output:
1750-01-01 1
1750-01-02 2
1750-01-03 3
...
1750-03-23 82
1750-03-24 83
1751-03-25 84
1751-03-26 85
...
1751-12-31 365
Chapter 18
Cheat sheet
This chapter explains how to perform some common — and some not so common — tasks in gretl’s
scripting language, hansl. Some but not all of the techniques listed here are also available through
the graphical interface. Although the graphical interface may be more intuitive and less intimidat-
ing at first, we encourage users to take advantage of the power of gretl’s scripting language as soon
as they feel comfortable with the program.
Comment: Now functions like sdiff() (“seasonal” difference) or estimation methods like seasonal
ARIMA will work as expected.
nulldata NT --preserve
setobs T 1:1 --stacked-time-series
Comment: The essential ingredient that we use here is the --preserve option: it protects existing
scalars (and matrices, for that matter) from being trashed by nulldata, thus making it possible to
use the scalar T in the setobs command.
setobs 1 1 --cross-section
series sortkey = -obs
dataset sortby sortkey
setobs 1 1950 --time-series
152
Chapter 18. Cheat sheet 153
Comment: The first line is required only if the data currently have a time series interpretation: it
removes that interpretation, because (for fairly obvious reasons) the dataset sortby operation is
not allowed for time series data. The following two lines reverse the data, using the negative of the
built-in index variable obs. The last line is just illustrative: it establishes the data as annual time
series, starting in 1950.
If you have a dataset that is mostly the right way round, but a particular variable is wrong, you can
reverse that variable as follows:
x = sortby(-obs, x)
list X = x1 x2 x3
smpl --no-missing X
Comment: You can now save the file via a store command to preserve a subsampled version of
the dataset. Alternative solutions based on the ok function, such as
list X = x1 x2 x3
series sel = ok(X)
smpl sel --restrict
are perhaps less obvious, but more flexible. Pick your poison.
“By” operations
Problem: You have a discrete variable d and you want to run some commands (for example, estimate
a model) by splitting the sample according to the values of d.
Solution:
matrix vd = values(d)
m = rows(vd)
loop i=1..m
scalar sel = vd[i]
smpl d==sel --restrict --replace
ols y const x
endloop
smpl --full
Comment: The main ingredient here is a loop. You can have gretl perform as many instructions as
you want for each value of d, as long as they are allowed inside a loop. Note, however, that if all
you want is descriptive statistics, the summary command does have a --by option.
date cpi
1990 130.658
1995 152.383
2000 172.192
What we need is for the CPI variable in the panel to repeat these three values 500 times.
Solution: Simple! With the panel dataset open in gretl,
append cpi.txt
Comment: If the length of the time series is the same as the length of the time dimension in the
panel (3 in this example), gretl will perform the stacking automatically. Rather than using the
append command you could use the “Append data” item under the File menu in the GUI program.
For this to work, your main dataset must be recognized as a panel. This can be arranged via the
setobs command or the “Dataset structure” item under the Data menu.
nulldata 36
set seed 61218
setobs 12 1:1 --stacked-time-series
###
### generate simulated yearly data
###
###
### now re-cast as 4-year averages
###
# compute averages
loop foreach i X --quiet
series $i = movavg($i, 4)
endloop
? print year X -o
year y x
1:01 2001 1 1
1:02 2002 -1 1
1:03 2003 -1 0
1:04 2004 0 1
1:05 2005 1 2
1:06 2006 1 2
1:07 2007 -1 0
1:08 2008 1 1
1:09 2009 0 3
1:10 2010 -1 1
1:11 2011 1 1
1:12 2012 0 1
...
3:09 2009 0 1
3:10 2010 1 1
3:11 2011 0 2
3:12 2012 1 2
? print id year X -o
id year y x
series d = (t=="1984:2")
Comment: The internal variable t is used to refer to observations in string form, so if you have a
cross-section sample you may just use d = (t=="123"). If the dataset has observation labels you
can use the corresponding label. For example, if you open the dataset mrw.gdt, supplied with gretl
among the examples, a dummy variable for Italy could be generated via
Note that this method does not require scripting at all. In fact, you might as well use the GUI Menu
“Add/Define new variable” for the same purpose, with the same syntax.
Chapter 18. Cheat sheet 156
Comment: Suppose you have a list D of mutually exclusive dummies, that is a full set of 0/1 vari-
ables coding for the value of some characteristic, such that the sum of the values of the elements
of D is 1 at each observation. This is, by the way, exactly what the dummify command produces.
The reverse job of dummify can be performed neatly by using the lincomb function.
The code above multiplies the first dummy variable in the list D by 1, the second one by 2, and so
on. Hence, the return value is a series whose value is i if and only if the i-th member of D has value
1.
If you want your coding to start from 0 instead of 1, you’ll have to modify the code snippet above
into
Easter
Problem: I have a 7-day daily dataset. How do I create an “Easter” dummy?
Solution: We have the easterday() function, which returns month and day of Easter given the
year. The following is an example script which uses this function and a few string magic tricks:
series Easter = 0
loop y=2011..2016
a = easterday(y)
m = floor(a)
d = round(100*(a-m))
ed_str = sprintf("%04d-%02d-%02d", y, m, d)
Easter["@ed_str"] = 1
endloop
Comment: The round() function is necessary for the “day” component because otherwise floating-
point problems may ensue. Try the year 2015, for example.
Recoding a variable
Problem: You want to perform a 1-to-1 recode on a variable. For example, consider tennis points:
you may have a variable x holding values 1 to 3 and you want to recode it to 15, 30, 40.
Solution 1:
Solution 2:
Comment: There are many equivalent ways to achieve the same effect, but for simple cases such
as this, the replace function is simple and transparent. If you don’t mind using matrices, scripts
using replace can also be remarkably compact. Note that replace also performs n-to-1 (“surjec-
tive”) replacements, such as
which would turn all entries equal to 2, 3, 5, 11, 22 or 33 to 1 and leave the other ones unchanged.
Comment: The above works fine if the number of distinct values in the source to be condensed into
each dummy variable is fairly small, but it becomes cumbersome if a single dummy must comprise
dozens of source values.
Clever solution:
Comment: The subset of values to be grouped together can be written out as a matrix relatively
compactly (first line). The magic that turns this into the desired series (second line) relies on the
versatility of the “dot” (element-wise) matrix operators. The expression “{src}” gets a column-
vector version of the input series—call this x — and “vec(sel)’” gets the input matrix as a row
vector, in case it’s a column vector or a matrix with both dimensions greater than 1 — call this s. If
x is n × 1 and s is 1 × m, the “.=” operator produces an n × m result, each element (i, j) of which
equals 1 if xi = sj , otherwise 0. The maxr() function along with the “.>” operator (see chapter 16
for both) then produces the result we want.
Of course, whichever procedure you use, you have to repeat for each of the dummy series you want
to create (but keep reading—the “proper” solution is probably what you want if you plan to create
several dummies).
Further comment: Note that the clever solution depends on converting what is “naturally” a vector
result into a series. This will fail if there are missing values in src, since (by default) missing values
will be skipped when converting src to x, and so the number of rows in the result will fall short
of the number of observations in the dataset. One fix is then to subsample the dataset to exclude
missing values before employing this method; another is to adjust the skip_missing setting via
the set command (see the Gretl Command Reference).
Proper solution:
The best solution, in terms of both computational efficiency and code clarity, would be using a
“conversion table” and the replace function, to produce a series on which the dummify command
can be used. For example, suppose we want to convert from a series called fips holding FIPS
Chapter 18. Cheat sheet 158
codes1 for the 50 US states plus the District of Columbia to a series holding codes for the four
standard US regions. We could create a 2 × 51 matrix — call it srmap— with the 51 FIPS codes on
the first row and the corresponding region codes on the second, and then do
Generating an ARMA(1,1)
Problem: Generate yt = 0.9yt−1 + εt − 0.5εt−1 , with εt ∼ NIID(0, 1).
Recommended solution:
alpha = 0.9
theta = -0.5
series y = filter(normal(), {1, theta}, alpha)
alpha = 0.9
theta = -0.5
series e = normal()
series y = 0
series y = alpha * y(-1) + e + theta * e(-1)
Comment: The filter function is specifically designed for this purpose so in most cases you’ll
want to take advantage of its speed and flexibility. That said, in some cases you may want to
generate the series in a manner which is more transparent (maybe for teaching purposes).
In the second solution, the statement series y = 0 is necessary because the next statement eval-
uates y recursively, so y[1] must be set. Note that you must use the keyword series here instead
of writing genr y = 0 or simply y = 0, to ensure that y is a series and not a scalar.
Solution:
Comment: True and false expressions are evaluated as 1 and 0 respectively, so they can be ma-
nipulated algebraically as any other number. The same result could also be achieved by using the
conditional assignment operator (see below), but in most cases it would probably lead to more
convoluted constructs.
Conditional assignment
Problem: Generate yt via the following rule:
(
xt for dt > a
yt =
zt for dt ≤ a
1 FIPS is the Federal Information Processing Standard: it assigns numeric codes from 1 to 56 to the US states and
outlying areas.
Chapter 18. Cheat sheet 159
Solution:
series y = (d > a) ? x : z
Comment: There are several alternatives to the one presented above. One is a brute force solution
using loops. Another one, more efficient but still suboptimal, would be
However, the ternary conditional assignment operator is not only the most numerically efficient
way to accomplish what we want, it is also remarkably transparent to read when one gets used to
it. Some readers may find it helpful to note that the conditional assignment operator works exactly
the same way as the =IF() function in spreadsheets.
series x = time
Comment: The special construct genr time and its variants are aware of whether a dataset is a
panel.
nulldata 20
x = normal()
y = normal()
z = x + y # collinear
list A = x y const z
list B = sanitize(A)
list print A
list print B
returns
? list print A
x y const z
? list print B
const x y
Chapter 18. Cheat sheet 160
Besides: it has been brought to our attention that some mischievous programs out there put the
constant last, instead of first, like God intended. We are not amused by this utter disrespect of
econometric tradition, but if you want to pursue the way of evil, it is rather simple to adapt the
script above to that effect.
# method 1
leverage --save --quiet
series h1 = lever
# method 2
series h2 = diag(qform(mX, iXX))
# method 3
series h3 = sumr(mX .* (mX*iXX))
# method 4
series h4 = NA
loop i=1..$nobs --quiet
matrix x = mX[i,]’
h4[i] = x’iXX*x
endloop
# verify
print h1 h2 h3 h4 --byobs
Comment: Solution 1 is the preferable one: it relies on the built-in leverage command, which
computes the requested series quite efficiently, taking care of missing values, possible restrictions
to the sample, etc.
However, three more are shown for didactical purposes, mainly to show the user how to manipulate
matrices. Solution 2 first constructs the PX matrix explicitly, via the qform function, and then takes
its diagonal; this is definitely not recommended (despite its compactness), since you generate a
much bigger matrix than you actually need and waste a lot of memory and CPU cycles in the
process. It doesn’t matter very much in the present case, since the sample size is very small, but
with a big dataset this could be a very bad idea.
Solution 3 is more clever, and relies on the fact that, if you define Z = X · (X 0 X)−1 , then hi could
also be written as
k
X
hi = x0i zi = xik zik
i=1
Chapter 18. Cheat sheet 161
which is in turn equivalent to the sum of the elements of the i-th row of X Z, where is the
element-by-element product. In this case, your clever usage of matrix algebra would produce a
solution computationally much superior to solution 2.
Solution 4 is the most old-fashioned one, and employs an indexed loop. While this wastes practi-
cally no memory and employs no more CPU cycles in algebraic operations than strictly necessary,
it imposes a much greater burden on the hansl interpreter, since handling a loop is conceptually
more complex than a single operation. In practice, you’ll find that for any realistically-sized prob-
lem, solution 4 is much slower that solution 3.
open bjg.gdt
order = 12
list L = lg || lags(order-1, lg)
smpl +order ;
series movmin = min(L)
series movmax = max(L)
series movmed = median(L)
smpl full
computes the moving minimum, maximum and median of the lg series. Plotting the four series
would produce something similar to figure 18.1.
6.6
lg
6.4 movmin
movmed
6.2 movmax
5.8
5.6
5.4
5.2
4.8
4.6
1950 1952 1954 1956 1958 1960
Solution: The Cholesky decomposition is your friend. If you want to generate data with a given
population covariance matrix, then all you have to do is post-multiply your pseudo-random data by
the Cholesky factor (transposed) of the matrix you want. For example:
? eval mcov(X)
2.0016 1.0157
1.0157 1.0306
If, instead, you want your simulated data to have a given sample covariance matrix, you have to
apply the same technique twice: one for standardizing the data, another one for giving it the
covariance structure you want. Example:
S = {2,1;1,1}
T = 1000
X = mnormal(T, rows(S))
X = X * (cholesky(S)/cholesky(mcov(X)))’
eval mcov(X)
gives you
? eval mcov(X)
2 1
1 1
as required.
list X = x1 x2 x3
list Z = z1 z2
list dZ = null
loop foreach i Z
series d$i = d * $i
list dZ = dZ d$i
endloop
ols y X Z d dZ
Comment: It’s amazing what string substitution can do for you, isn’t it?
Chapter 18. Cheat sheet 163
Realized volatility
Given data by the minute, you want to compute the “realized volatility” for the hour as
Problem: P
1 60 2
RVt = 60 τ=1 yt:τ . Imagine your sample starts at time 1:1.
Solution:
smpl --full
genr time
series minute = int(time/60) + 1
series second = time % 60
setobs minute second --panel
series rv = psd(y)^2
setobs 1 1
smpl second==1 --restrict
store foo rv
Comment: Here we trick gretl into thinking that our dataset is a panel dataset, where the minutes
are the “units” and the seconds are the “time”; this way, we can take advantage of the special
function psd(), panel standard deviation. Then we simply drop all observations but one per minute
and save the resulting data (store foo rv translates as “store in the gretl datafile foo.gdt the
series rv”).
list L1 = a b c
list L2 = x y z
k1 = 1
loop foreach i L1 --quiet
k2 = 1
loop foreach j L2 --quiet
if k1 == k2
ols $i 0 $j
endif
k2++
endloop
k1++
endloop
Comment: The simplest way to achieve the result is to loop over all possible combinations and
filter out the unneeded ones via an if condition, as above. That said, in some cases variable names
can help. For example, if
list Lx = x1 x2 x3
list Ly = y1 y2 y3
then we could just loop over the integers — quite intuitive and certainly more elegant:
loop i=1..3
ols y$i const x$i
endloop
Chapter 18. Cheat sheet 164
where
k
X
rk = pi qk−i
i=0
is the convolution of the pi and qi coefficients. The same operation can be performed via the FFT,
but in most cases using conv2d is quicker and more natural.
As an example, we’ll use the same one we used in Section 9.5: consider the multiplication of two
polynomials:
P (x) = 1 + 0.5x
Q(x) = 1 + 0.3x − 0.8x 2
R(x) = P (x) · Q(x) = 1 + 0.8x − 0.65x 2 − 0.4x 3
p = {1; 0.5}
q = {1; 0.3; -0.8}
r = conv2d(p, q)
print r
r (4 x 1)
1
0.8
-0.65
-0.4
which is indeed the desired result. Note that the same computation could also be performed via
the filter function, at the price of slightly more elaborate syntax.
A 4 B = (A \ B) ∪ (B \ A)
where in this context backslash represents the relative complement operator, such that
A \ B = {x ∈ A | x 6∈ B}
Chapter 18. Cheat sheet 165
In practice we first check if there are any series in A but not in B, then we perform the reverse check.
If the union of the two results is an empty set, then the lists must contain the same variables. The
hansl syntax for this would be something like
list X print
list revX print
will produce
? list X print
D1 D872 EEN_DIS GCD
? list revX print
GCD EEN_DIS D872 D1
Cross-validation
Problem: “I’d like to compute the so-called leave-one-out cross-validation criterion for my regression.
Is there a command in gretl?”
If you have a sample with n observations, the “leave-one-out” cross-validation criterion can be
mechanically computed by running n regressions in which one observation at a time is omitted
and all the other ones are used to forecast its value. The sum of the n squared forecast errors is
the statistic we want. Fortunately, there is no need to do so. It is possible to prove that the same
statistic can be computed as
Xn
CV = [ûi /(1 − hi )]2 ,
i=1
where hi is the i-th element of the “hat” matrix (see section 18.2) from a regression on the whole
sample.
Chapter 18. Cheat sheet 166
This method is natively provided by gretl as a side benefit to the leverage command, that stores
the CV criterion into the $test accessor. The following script shows the equivalence of the two
approaches:
scalar CV = 0
matrix mX = {X}
loop i = 1 .. $nobs
xi = mX[i,]
yi = price[i]
smpl obs != i --restrict
ols price X --quiet
smpl full
scalar fe = yi - xi * $coeff
CV += fe^2
endloop
sumc(sumr(!ok(m))) > 0
If this gives a non-zero return value you know that m contains at least one non-finite element.
Part II
Econometric methods
167
Chapter 19
19.1 Introduction
Consider (once again) the linear regression model
y = Xβ + u (19.1)
β̂ = (X 0 X)−1 X 0 y (19.2)
If the condition E(u|X) = 0 is satisfied, this is an unbiased estimator; under somewhat weaker
conditions the estimator is biased but consistent. It is straightforward to show that when the OLS
estimator is unbiased (that is, when E(β̂ − β) = 0), its variance is
Var(β̂) = E (β̂ − β)(β̂ − β)0 = (X 0 X)−1 X 0 ΩX(X 0 X)−1 (19.3)
If the iid assumption is not satisfied, two things follow. First, it is possible in principle to construct
a more efficient estimator than OLS—for instance some sort of Feasible Generalized Least Squares
(FGLS). Second, the simple “classical” formula for the variance of the least squares estimator is no
longer correct, and hence the conventional OLS standard errors — which are just the square roots
of the diagonal elements of the matrix defined by (19.4) — do not provide valid means of statistical
inference.
In the recent history of econometrics there are broadly two approaches to the problem of non-
iid errors. The “traditional” approach is to use an FGLS estimator. For example, if the departure
from the iid condition takes the form of time-series dependence, and if one believes that this
could be modeled as a case of first-order autocorrelation, one might employ an AR(1) estimation
method such as Cochrane–Orcutt, Hildreth–Lu, or Prais–Winsten. If the problem is that the error
variance is non-constant across observations, one might estimate the variance as a function of the
independent variables and then perform weighted least squares, using as weights the reciprocals
of the estimated variances.
While these methods are still in use, an alternative approach has found increasing favor: that is,
use OLS but compute standard errors (or more generally, covariance matrices) that are robust with
respect to deviations from the iid assumption. This is typically combined with an emphasis on
using large datasets—large enough that the researcher can place some reliance on the (asymptotic)
consistency property of OLS. This approach has been enabled by the availability of cheap computing
power. The computation of robust standard errors and the handling of very large datasets were
daunting tasks at one time, but now they are unproblematic. The other point favoring the newer
methodology is that while FGLS offers an efficiency advantage in principle, it often involves making
168
Chapter 19. Robust covariance matrix estimation 169
additional statistical assumptions which may or may not be justified, which may not be easy to test
rigorously, and which may threaten the consistency of the estimator — for example, the “common
factor restriction” that is implied by traditional FGLS “corrections” for autocorrelated errors.
James Stock and Mark Watson’s Introduction to Econometrics illustrates this approach at the level of
undergraduate instruction: many of the datasets they use comprise thousands or tens of thousands
of observations; FGLS is downplayed; and robust standard errors are reported as a matter of course.
In fact, the discussion of the classical standard errors (labeled “homoskedasticity-only”) is confined
to an Appendix.
Against this background it may be useful to set out and discuss all the various options offered
by gretl in respect of robust covariance matrix estimation. The first point to notice is that gretl
produces “classical” standard errors by default (in all cases apart from GMM estimation). In script
mode you can get robust standard errors by appending the --robust flag to estimation commands.
In the GUI program the model specification dialog usually contains a “Robust standard errors”
check box, along with a “configure” button that is activated when the box is checked. The configure
button takes you to a configuration dialog (which can also be reached from the main menu bar:
Tools → Preferences → General → HCCME). There you can select from a set of possible robust
estimation variants, and can also choose to make robust estimation the default.
The specifics of the available options depend on the nature of the data under consideration —
cross-sectional, time series or panel— and also to some extent the choice of estimator. (Although
we introduced robust standard errors in the context of OLS above, they may be used in conjunction
with other estimators too.) The following three sections of this chapter deal with matters that are
specific to the three sorts of data just mentioned. Note that additional details regarding covariance
matrix estimation in the context of GMM are given in chapter 24.
We close this introduction with a brief statement of what “robust standard errors” can and cannot
achieve. They can provide for asymptotically valid statistical inference in models that are basically
correctly specified, but in which the errors are not iid. The “asymptotic” part means that they
may be of little use in small samples. The “correct specification” part means that they are not a
magic bullet: if the error term is correlated with the regressors, so that the parameter estimates
themselves are biased and inconsistent, robust standard errors will not save the day.
1 In some specialized contexts spatial autocorrelation may be an issue. Gretl does not have any built-in methods to
small” on average. This point is quite intuitive. The OLS parameter estimates, β̂, satisfy by design
the criterion that the sum of squared residuals,
X X 2
û2t = yt − Xt β̂
is minimized for given X and y. Suppose that β̂ ≠ β. This is almost certain to be the case: even if
OLS is not biased, it would be a miracle if the β̂ calculated from any finite sample were
P exactly equal
to β. But in that case thePsum of squares of the true, unobserved errors, u2t = (yt − Xt β)2 is
P
bound to be greater than û2t . The elaborated variants on HC0 take this point on board as follows:
• HC1 : Applies a degrees-of-freedom correction, multiplying the HC0 matrix by T /(T − k).
• HC2 : Instead of using û2t for the diagonal elements of Ω̂, uses û2t /(1 − ht ), where ht =
Xt (X 0 X)−1 Xt0 , the t th diagonal element of the projection matrix, P , which has the property
that P · y = ŷ. The relevance of ht is that if the variance of all the ut is σ 2 , the expectation
of û2t is σ 2 (1 − ht ), or in other words, the ratio û2t /(1 − ht ) has expectation σ 2 . As Davidson
and MacKinnon show, 0 ≤ ht < 1 for all t, so this adjustment cannot reduce the the diagonal
elements of Ω̂ and in general revises them upward.
• HC3 : Uses û2t /(1 − ht )2 . The additional factor of (1 − ht ) in the denominator, relative to
HC2 , may be justified on the grounds that observations with large variances tend to exert a
lot of influence on the OLS estimates, so that the corresponding residuals tend to be under-
estimated. See Davidson and MacKinnon for a fuller explanation.
The relative merits of these variants have been explored by means of both simulations and the-
oretical analysis. Unfortunately there is not a clear consensus on which is “best”. Davidson and
MacKinnon argue that the original HC0 is likely to perform worse than the others; nonetheless,
“White’s standard errors” are reported more often than the more sophisticated variants and there-
fore, for reasons of comparability, HC0 is the default HCCME in gretl.
If you wish to use HC1 , HC2 or HC3 you can arrange for this in either of two ways. In script mode,
you can do, for example,
set hc_version 2
In the GUI program you can go to the HCCME configuration dialog, as noted above, and choose any
of these variants to be the default.
Despite the points made above, some residual degree of heteroskedasticity may be present in time
series data: the key point is that in most cases it is likely to be combined with serial correlation
(autocorrelation), hence demanding a special treatment. In White’s approach, Ω̂, the estimated
covariance matrix of the ut , remains conveniently diagonal: the variances, E(u2t ), may differ by
t but the covariances, E(ut us ), are all zero. Autocorrelation in time series data means that at
least some of the the off-diagonal elements of Ω̂ should be non-zero. This introduces a substantial
complication and requires another piece of terminology; estimates of the covariance matrix that
are asymptotically valid in face of both heteroskedasticity and autocorrelation of the error process
are termed HAC (heteroskedasticity and autocorrelation consistent).
The issue of HAC estimation is treated in more technical terms in chapter 24. Here we try to
convey some of the intuition at a more basic level. We begin with a general comment: residual
autocorrelation is not so much a property of the data, as a symptom of an inadequate model. Data
may be persistent though time, and if we fit a model that does not take this aspect into account
properly, we end up with a model with autocorrelated disturbances. Conversely, it is often possible
to mitigate or even eliminate the problem of autocorrelation by including relevant lagged variables
in a time series model, or in other words, by specifying the dynamics of the model more fully. HAC
estimation should not be seen as the first resort in dealing with an autocorrelated error process.
That said, the “obvious” extension of White’s HCCME to the case of autocorrelated errors would
seem to be this: estimate the off-diagonal elements of Ω̂ (that is, the autocovariances, E(ut us ))
using, once again, the appropriate OLS residuals: ω̂ts = ût ûs . This is basically right, but demands
an important amendment. We seek a consistent estimator, one that converges towards the true Ω
as the sample size tends towards infinity. This can’t work if we allow unbounded serial depen-
dence. Bigger samples will enable us to estimate more of the true ωts elements (that is, for t and
s more widely separated in time) but will not contribute ever-increasing information regarding the
maximally separated ωts pairs, since the maximal separation itself grows with the sample size.
To ensure consistency, we have to confine our attention to processes exhibiting temporally limited
dependence, or in other words cut off the computation of the ω̂ts values at some maximum value
of p = t − s (where p is treated as an increasing function of the sample size, T , although it cannot
increase in proportion to T ).
The simplest variant of this idea is to truncate the computation at some finite lag order p, where
p grows as, say, T 1/4 . The trouble with this is that the resulting Ω̂ may not be a positive definite
matrix. In practical terms, we may end up with negative estimated variances. One solution to this
problem is offered by The Newey–West estimator (Newey and West, 1987), which assigns declining
weights to the sample autocovariances as the temporal separation increases.
To understand this point it is helpful to look more closely at the covariance matrix given in (19.5),
namely,
(X 0 X)−1 (X 0 Ω̂X)(X 0 X)−1
This is known as a “sandwich” estimator. The bread, which appears on both sides, is (X 0 X)−1 .
This is a k × k matrix, and is also the key ingredient in the computation of the classical covariance
matrix. The filling in the sandwich is
Σ̂ = X0 Ω̂ X
(k×k) (k×T ) (T ×T ) (T ×k)
It can be proven that, under mild regularity conditions, T −1 Σ̂ is a consistent estimator of the long-
run covariance matrix of the random k-vector xt · ut .
From a computational point of view, it is neither necessary nor desirable to store the (potentially
very large) T × T matrix Ω̂ as such. Rather, one computes the sandwich filling by summation as a
weighted sum:
p
X
Σ̂ = Γ̂ (0) + wj Γ̂ (j) + Γ̂ 0 (j)
j=1
Chapter 19. Robust covariance matrix estimation 172
where wj is the weight given to lag j > 0 and the k × k matrix Γ̂ (j), for j ≥ 0, is given by
T
X
Γ̂ (j) = ût ût−j Xt0 Xt−j ;
t=j+1
that is, the sample autocovariance matrix of xt · ut at lag j, apart from a scaling factor T .
This leaves two questions. How exactly do we determine the maximum lag length or “bandwidth”,
p, of the HAC estimator? And how exactly are the weights wj to be determined? We will return to
the (difficult) question of the bandwidth shortly. As regards the weights, gretl offers three variants.
The default is the Bartlett kernel, as used by Newey and West. This sets
1− j j≤p
p+1
wj =
0 j>p
so the weights decline linearly as j increases. The other two options are the Parzen kernel and the
Quadratic Spectral (QS) kernel. For the Parzen kernel,
3
2
1 − 6aj + 6aj 0 ≤ aj ≤ 0.5
wj = 2(1 − aj )3 0.5 < aj ≤ 1
0 a >1
j
Bartlett Parzen QS
In gretl you select the kernel using the set command with the hac_kernel parameter:
As shown in Table 19.1 the choice between nw1 and nw2 does not make a great deal of difference.
T p (nw1) p (nw2)
50 2 3
100 3 4
150 3 4
200 4 4
300 5 5
400 5 5
You also have the option of specifying a fixed numerical value for p, as in
set hac_lag 6
In addition you can set a distinct bandwidth for use with the Quadratic Spectral kernel (since this
need not be an integer). For example,
VLR (ut )
VLR (xt ) =
(1 − ρ)2
In most cases, ut is likely to be less autocorrelated than xt , so a smaller bandwidth should suffice.
Estimation of VLR (xt ) can therefore proceed in three steps: (1) estimate ρ; (2) obtain a HAC estimate
of ût = xt − ρ̂xt−1 ; and (3) divide the result by (1 − ρ)2 .
The application of the above concept to our problem implies estimating a finite-order Vector Au-
toregression (VAR) on the vector variables ξt = Xt ût . In general, the VAR can be of any order, but
in most cases 1 is sufficient; the aim is not to build a watertight model for ξt , but just to “mop up”
a substantial part of the autocorrelation. Hence, the following VAR is estimated
ξt = Aξt−1 + εt
set hac_prewhiten on
Chapter 19. Robust covariance matrix estimation 174
There is at present no mechanism for specifying an order other than 1 for the initial VAR.
A further refinement is available in this context, namely data-based bandwidth selection. It makes
intuitive sense that the HAC bandwidth should not simply be based on the size of the sample,
but should somehow take into account the time-series properties of the data (and also the kernel
chosen). A nonparametric method for doing this was proposed by Newey and West (1994); a good
concise account of the method is given in Hall (2005). This option can be invoked in gretl via
This option is the default when prewhitening is selected, but you can override it by giving a specific
numerical value for hac_lag.
Even the Newey–West data-based method does not fully pin down the bandwidth for any particular
sample. The first step involves calculating a series of residual covariances. The length of this series
is given as a function of the sample size, but only up to a scalar multiple — for example, it is given
as O(T 2/9 ) for the Bartlett kernel. Gretl uses an implied multiple of 1.
• The variance of the error term may differ across the cross-sectional units.
• The covariance of the errors across the units may be non-zero in each time period.
• If the “between” variation is not swept out, the errors may exhibit autocorrelation, not in the
usual time-series sense but in the sense that the mean value of the error term may differ
across units. This is relevant when estimation is by pooled OLS.
Gretl currently offers two robust covariance matrix estimators specifically for panel data. These are
available for models estimated via fixed effects, random effects, pooled OLS, and pooled two-stage
least squares. The default robust estimator is that suggested by Arellano (2003), which is HAC
provided the panel is of the “large n, small T ” variety (that is, many units are observed in relatively
few periods). The Arellano estimator is
n
−1 −1
X
Σ̂A = X 0 X Xi0 ûi û0i Xi X 0 X
i=1
where X is the matrix of regressors (with the group means subtracted in the case of fixed effects, or
quasi-demeaned in the case of random effects) ûi denotes the vector of residuals for unit i, and n
is the number of cross-sectional units.2 Cameron and Trivedi (2005) make a strong case for using
this estimator; they note that the ordinary White HCCME can produce misleadingly small standard
2 This variance estimator is also known as the “clustered (over entities)” estimator.
Chapter 19. Robust covariance matrix estimation 175
errors in the panel context because it fails to take autocorrelation into account.3 In addition Stock
and Watson (2008) show that the White HCCME is inconsistent in the fixed-effects panel context for
fixed T > 2.
In cases where autocorrelation is not an issue the estimator proposed by Beck and Katz (1995)
and discussed by Greene (2003, chapter 13) may be appropriate. This estimator, which takes into
account contemporaneous correlation across the units and heteroskedasticity by unit, is
n Xn
0
−1 X 0
−1
Σ̂BK = X X σ̂ij Xi Xj X 0 X
i=1 j=1
set pcse on
(Note that regardless of the pcse setting, the robust estimator is not used unless the --robust flag
is given, or the “Robust” box is checked in the GUI program.)
where m denotes the number of clusters, and Xj and ûj denote, respectively, the matrix of regres-
sors and the vector of residuals that fall within cluster j. As noted above, the Arellano variance
estimator for panel data models is a special case of this, where the clustering is by panel unit.
For models estimated by the method of Maximum Likelihood (in which case the standard variance
estimator is the inverse of the negative Hessian, H), the cluster estimator is
m
X
Σ̂C = H −1 Gj0 Gj H −1
j=1
where Gj is the sum of the “score” (that is, the derivative of the loglikelihood with respect to the
parameter estimates) across the observations falling within cluster j.
3 See also Cameron and Miller (2015) for a discussion of the Arellano-type estimator in the context of the random
effects model.
Chapter 19. Robust covariance matrix estimation 176
It is common to apply a degrees of freedom adjustment to these estimators (otherwise the variance
may appear misleadingly small in comparison with other estimators, if the number of clusters is
small). In the least squares case the factor is (m/(m − 1)) × (n − 1)/(n − k), where n is the total
number of observations and k is the number of parameters estimated; in the case of ML estimation
the factor is just m/(m − 1).
ols y 0 x1 x2 --cluster=cvar
The specified clustering variable must (a) be defined (not missing) at all observations used in esti-
mating the model and (b) take on at least two distinct values over the estimation range. The clusters
are defined as sets of observations having a common value for the clustering variable. It is generally
expected that the number of clusters is substantially less than the total number of observations.
Chapter 20
Panel data
A panel dataset is one in which each of N > 1 units (sometimes called “individuals” or “groups”) is
observed over time. In a balanced panel there are T > 1 observations on each unit; more generally
the number of observations may differ by unit. In the following we index units by i and time by t.
To allow for imbalance in a panel we use the notation Ti to refer to the number of observations for
unit or individual i.
where yit is the observation on the dependent variable for cross-sectional unit i in period t, Xit
is a 1 × k vector of independent variables observed for unit i in period t, β is a k × 1 vector of
parameters, and uit is an error or disturbance term specific to unit i in period t.
The fixed and random effects models have in common that they decompose the unitary pooled
error term, uit . For the fixed effects model we write uit = αi + εit , yielding
That is, we decompose uit into a unit-specific and time-invariant component, αi , and an observation-
specific error, εit .1 The αi s are then treated as fixed parameters (in effect, unit-specific y-intercepts),
which are to be estimated. This can be done by including a dummy variable for each cross-sectional
1 It is possible to break a third component out of u , namely w , a shock that is time-specific but common to all the
it t
units in a given period. In the interest of simplicity we do not pursue that option here.
177
Chapter 20. Panel data 178
unit (and suppressing the global constant). This is sometimes called the Least Squares Dummy Vari-
ables (LSDV) method. Alternatively, one can subtract the group mean from each of variables and
estimate a model without a constant. In the latter case the dependent variable may be written as
where Ti is the number of observations for unit i. An exactly analogous formulation applies to the
independent variables. Given parameter estimates, β̂, obtained using such de-meaned data we can
recover estimates of the αi s using
Ti
1 X
α̂i = yit − Xit β̂
Ti t=1
These two methods (LSDV, and using de-meaned data) are numerically equivalent. gretl takes the
approach of de-meaning the data. If you have a small number of cross-sectional units, a large num-
ber of time-series observations per unit, and a large number of regressors, it is more economical
in terms of computer memory to use LSDV. If need be you can easily implement this manually. For
example,
genr unitdum
ols y x du_*
In contrast to the fixed effects model, the vi s are not treated as fixed parameters, but as random
drawings from a given probability distribution.
The celebrated Gauss–Markov theorem, according to which OLS is the best linear unbiased esti-
mator (BLUE), depends on the assumption that the error term is independently and identically
distributed (IID). In the panel context, the IID assumption means that E(u2it ), in relation to equa-
tion 20.1, equals a constant, σu2 , for all i and t, while the covariance E(uis uit ) equals zero for all
s ≠ t and the covariance E(ujt uit ) equals zero for all j ≠ i.
If these assumptions are not met—and they are unlikely to be met in the context of panel data —
OLS is not the most efficient estimator. Greater efficiency may be gained using generalized least
squares (GLS), taking into account the covariance structure of the error term.
Consider observations on a given unit i at two different times s and t. From the hypotheses above
it can be worked out that Var(uis ) = Var(uit ) = σv2 + σε2 , while the covariance between uis and uit
is given by E(uis uit ) = σv2 .
In matrix notation, we may group all the Ti observations for unit i into the vector yi and write it as
yi = Xi β + ui (20.4)
Chapter 20. Panel data 179
The vector ui , which includes all the disturbances for individual i, has a variance–covariance matrix
given by
Var(ui ) = Σi = σε2 I + σv2 J (20.5)
where J is a square matrix with all elements equal to 1. It can be shown that the matrix
θi
Ki = I − J,
Ti
r
where θi = 1 − σε2 / σε2 + Ti σv2 , has the property
Ki ΣKi0 = σε2 I
Ki yi = Ki Xi β + Ki ui (20.6)
satisfies the Gauss–Markov conditions, and OLS estimation of (20.6) provides efficient inference.
But since
Ki yi = yi − θi ȳi
GLS estimation is equivalent to OLS using “quasi-demeaned” variables; that is, variables from which
we subtract a fraction θ of their average.2 Notice that for σε2 → 0, θ → 1, while for σv2 → 0, θ → 0.
This means that if all the variance is attributable to the individual effects, then the fixed effects
estimator is optimal; if, on the other hand, individual effects are negligible, then pooled OLS turns
out, unsurprisingly, to be the optimal estimator.
To implement the GLS approach we need to calculate θ, which in turn requires estimates of the
two variances σε2 and σv2 . (These are often referred to as the “within” and “between” variances
respectively, since the former refers to variation within each cross-sectional unit and the latter to
variation between the units). Several means of estimating these magnitudes have been suggested in
the literature (see Baltagi, 1995); by default gretl uses the method of Swamy and Arora (1972): σε2
is estimated by the residual variance from the fixed effects model, and σv2 is estimated indirectly
with the help of the “between” regression which uses the group means of all the relevant variables:
is,
ȳi = X̄i β + ei
The residual variance from this regression, se2 , can be shown to estimate the sum σv2 + σε2 /T . An
estimate of σv2 can therefore be obtained by subtracting 1/T times the estimate of σε2 from se2 :
Alternatively, if the --nerlove option is given, gretl uses the method suggested by Nerlove (1971).
In this case σv2 is estimated as the sample variance of the fixed effects, α̂i ,
n
1 X ¯
2
σ̂v2 = α̂i − α̂ (20.8)
N − 1 i=1
option with the panel command. But the gain in efficiency from doing so may well be slim; for a
discussion of this point and related matters see Cottrell (2017). Unbalancedness also affects the
Nerlove (1971) estimator, but the econometric literature offers no guidance on the details. Gretl
uses the weighted average of the fixed effects as a natural extension of the original method. Again,
see Cottrell (2017) for further details.
Choice of estimator
Which panel method should one use, fixed effects or random effects?
One way of answering this question is in relation to the nature of the data set. If the panel comprises
observations on a fixed and relatively small set of units of interest (say, the member states of the
European Union), there is a presumption in favor of fixed effects. If it comprises observations on a
large number of randomly selected individuals (as in many epidemiological and other longitudinal
studies), there is a presumption in favor of random effects.
Besides this general heuristic, however, various statistical issues must be taken into account.
1. Some panel data sets contain variables whose values are specific to the cross-sectional unit
but which do not vary over time. If you want to include such variables in the model, the fixed
effects option is simply not available. When the fixed effects approach is implemented using
dummy variables, the problem is that the time-invariant variables are perfectly collinear with
the per-unit dummies. When using the approach of subtracting the group means, the issue is
that after de-meaning these variables are nothing but zeros.
2. A somewhat analogous issue arises with the random effects estimator. As mentioned above,
the default Swamy–Arora method relies on the group-means regression to obtain a measure
of the between variance. Suppose we have observations on n units or individuals and there
are k independent variables of interest. If k > n, this regression cannot be run — since we
have only n effective observations — and hence Swamy–Arora estimates cannot be obtained.
In this case, however, it is possible to use Nerlove’s method instead.
If both fixed effects and random effects are feasible for a given specification and dataset, the choice
between these estimators may be expressed in terms of the two econometric desiderata, efficiency
and consistency.
From a purely statistical viewpoint, we could say that there is a tradeoff between robustness and
efficiency. In the fixed effects approach, we do not make any hypotheses on the “group effects” (that
is, the time-invariant differences in mean between the groups) beyond the fact that they exist —
and that can be tested; see below. As a consequence, once these effects are swept out by taking
deviations from the group means, the remaining parameters can be estimated.
On the other hand, the random effects approach attempts to model the group effects as drawings
from a probability distribution instead of removing them. This requires that individual effects are
representable as a legitimate part of the disturbance term, that is, zero-mean random variables,
uncorrelated with the regressors.
As a consequence, the fixed-effects estimator “always works”, but at the cost of not being able to
estimate the effect of time-invariant regressors. The richer hypothesis set of the random-effects
estimator ensures that parameters for time-invariant regressors can be estimated, and that esti-
mation of the parameters for time-varying regressors is carried out more efficiently. These advan-
tages, though, are tied to the validity of the additional hypotheses. If, for example, there is reason
to think that individual effects may be correlated with some of the explanatory variables, then the
random-effects estimator would be inconsistent, while fixed-effects estimates would still be valid.
It is precisely on this principle that the Hausman test is built (see below): if the fixed- and random-
effects estimates agree, to within the usual statistical margin of error, there is no reason to think
the additional hypotheses invalid, and as a consequence, no reason not to use the more efficient RE
estimator.
Chapter 20. Panel data 181
• Collect the fixed-effects estimates in a vector β̃ and the corresponding random-effects esti-
mates in β̂, then form the difference vector (β̃ − β̂).
• Form the covariance matrix of the difference vector as Var(β̃ − β̂) = Var(β̃) − Var(β̂) = Ψ ,
where Var(β̃) and Var(β̂) are estimated by the sample variance matrices of the fixed- and
random-effects models respectively.3
0
• Compute H = β̃ − β̂ Ψ −1 β̃ − β̂ .
Given the relative efficiencies of β̃ and β̂, the matrix Ψ “should be” positive definite, in which case
H is positive, but in finite samples this is not guaranteed and of course a negative χ 2 value is not
admissible. The regression method avoids this potential problem. The procedure is:
• Treat the random-effects model as the restricted model, and record its sum of squared resid-
uals as SSRr .
• Estimate via OLS an unrestricted model in which the dependent variable is quasi-demeaned y
and the regressors include both quasi-demeaned X (as in the RE model) and the de-meaned
variants of all the time-varying variables (i.e. the fixed-effects regressors); record the sum of
squared residuals from this model as SSRu .
3 Hausman (1978) showed that the covariance of the difference takes this simple form when β̂ is an efficient estimator
and β̃ is inefficient.
Chapter 20. Panel data 182
• Compute H = n (SSRr − SSRu ) /SSRu , where n is the total number of observations used. On
this variant H cannot be negative, since adding additional regressors to the RE model cannot
raise the SSR.
By default gretl computes the Hausman test via the regression method, but it uses the matrix-
difference method if you pass the option --matrix-diff to the panel command.
Serial correlation
A simple test for first-order autocorrelation of the error term, namely the Durbin–Watson (DW)
statistic, is printed as part of the output for pooled OLS as well as fixed-effects and random-effects
estimation. Let us define “serial correlation proper” as serial correlation strictly in the time di-
mension of a panel dataset. When based on the residuals from fixed-effects estimation, the DW
statistic is a test for serial correlation proper.4 The DW value shown in the case of random-effects
estimation is based on the fixed-effects residuals. When DW is based on pooled OLS residuals it
tests for serial correlation proper only on the assumption of a common intercept. Put differently,
in this case it tests a joint null hypothesis: absence of fixed effects plus absence of (first order)
serial correlation proper. In the presence of missing observations the DW statistic is calculated as
described in Baltagi and Wu (1999) (their expression for d1 under equation (16) on page 819).
When it is computed, the DW statistic can be retrieved via the accessor $dw after estimation. In
addition, an approximate P -value for the null of no serial correlation (ρ = 0) against the alternative
of ρ > 0 may be available via the accessor $dwpval. This is based on the analysis in Bhargava
et al. (1982); strictly speaking it is the marginal significance level of DW considered as a dL value
(the value below which the test rejects, as opposed to dU , the value above which the test fails to
reject). In the panel case, however, dL and dU are quite close, particularly when N (the number of
individual units) is large. At present gretl does not attempt to compute such P -values when the
number of observations differs across individuals.
open abdata.gdt
panel n const w k ys --fixed-effects
prediction yielded by the model is a desirable descriptive statistic to have, but which model and
which (variant of the) dependent variable are we talking about?
Fixed-effects models can be thought of in two equally defensible ways. From one perspective they
provide a nice, clean way of sweeping out individual effects by using the fact that in the linear
model a sufficient statistic is easy to compute. Alternatively, they provide a clever way to estimate
the “important” parameters of a model in which you want to include (for whatever reason) a full
set of individual dummies. If you take the second of these perspectives, your dependent variable is
unmodified y and your model includes the unit dummies; the appropriate R 2 measure is then the
squared correlation between y and the ŷ computed using both the measured individual effects and
the effects of the explicitly named regressors. This is reported by gretl as the “LSDV R-squared”. If
you take the first point of view, on the other hand, your dependent variable is really yit − ȳi and
your model just includes the β terms, the coefficients of deviations of the x variables from their
per-unit means. In this case, the relevant measure of R 2 is the so-called “within” R 2 ; this variant
is printed by gretl for fixed-effects model in place of the adjusted R 2 (it being unclear in this case
what exactly the “adjustment” should amount to anyway).
In this model gretl takes the “fitted value” ($yhat) to be α̂i + Xit β̂, and the residual ($uhat) to be
yit minus this fitted value. This makes sense because the fixed effects (the αi terms) are taken as
parameters to be estimated. However, it can be argued that the fixed effects are not really “explana-
tory” and if one defines the residual as the observed yit value minus its “explained” component
one might prefer to see just yit − Xit β̂. You can get this after fixed-effects estimation as follows:
where $ahat gives the unit-specific intercept (as it would be calculated if one included all N unit
dummies and omitted a common y-intercept), and $coeff[1] gives the “global” y-intercept.6
6 For anyone used to Stata, gretl’s fixed-effects $uhat corresponds to what you get from Stata’s “predict, e” after
In this case gretl considers the error term to be vi + εit (since vi is conceived as a random drawing)
and the $uhat series is an estimate of this, namely
yit − Xit β̂
What if you want an estimate of just vi (or just εit ) in this case? This poses a signal-extraction
problem: given the composite residual, how to recover an estimate of its components? The solution
is to ascribe to the individual effect, v̂i , a suitable fraction of the mean residual per individual,
¯ i = Ti ûit . The “suitable fraction” is the proportion of the variance of the variance of ūi that is
P
û t=1
due to vi , namely
σv2
= 1 − (1 − θi )2
σv + σε2 /Ti
2
After random effects estimation in gretl you can access a series containing the v̂i s under the name
$ahat. This series can be calculated by hand as follows:
First, if the error uit includes a group effect, vi , then yit−1 is bound to be correlated with the error,
since the value of vi affects yi at all t. That means that OLS applied to (20.9) will be inconsistent
as well as inefficient. The fixed-effects model sweeps out the group effects and so overcomes this
particular problem, but a subtler issue remains, which applies to both fixed and random effects
estimation. Consider the de-meaned representation of fixed effects, as applied to the dynamic
model,
ỹit = X̃it β + ρ ỹi,t−1 + εit
where ỹit = yit − ȳi and εit = uit − ūi (or uit − αi , using the notation of equation 20.2). The trouble
is that ỹi,t−1 will be correlated with εit via the group mean, ȳi . The disturbance εit influences yit
directly, which influences ȳi , which, by construction, affects the value of ỹit for all t. The same
issue arises in relation to the quasi-demeaning used for random effects. Estimators which ignore
this correlation will be consistent only as T → ∞ (in which case the marginal effect of εit on the
group mean of y tends to vanish).
One strategy for handling this problem, and producing consistent estimates of β and ρ, was pro-
posed by Anderson and Hsiao (1981). Instead of de-meaning the data, they suggest taking the first
difference of (20.9), an alternative tactic for sweeping out the group effects:
where ηit = ∆uit = ∆(vi + εit ) = εit − εi,t−1 . We’re not in the clear yet, given the structure of the
error ηit : the disturbance εi,t−1 is an influence on both ηit and ∆yi,t−1 = yit − yi,t−1 . The next step
is then to find an instrument for the “contaminated” ∆yi,t−1 . Anderson and Hsiao suggest using
either yi,t−2 or ∆yi,t−2 , both of which will be uncorrelated with ηit provided that the underlying
errors, εit , are not themselves serially correlated.
The Anderson–Hsiao estimator is not provided as a built-in function in gretl, since gretl’s sensible
handling of lags and differences for panel data makes it a simple application of regression with
instrumental variables—see Listing 20.2, which is based on a study of country growth rates by
Nerlove (1999).7
Although the Anderson–Hsiao estimator is consistent, it is not most efficient: it does not make the
fullest use of the available instruments for ∆yi,t−1 , nor does it take into account the differenced
structure of the error ηit . It is improved upon by the methods of Arellano and Bond (1991) and
Blundell and Bond (1998). These methods are taken up in the next chapter.
The primary command for estimating dynamic panel models in gretl is dpanel. The closely re-
lated arbond command predated dpanel, and is still present, but whereas arbond only supports
the so-called “difference” estimator (Arellano and Bond, 1991), dpanel in addition offers the “sys-
tem” estimator (Blundell and Bond, 1998), which has become the method of choice in the applied
literature.
21.1 Introduction
Notation
A dynamic linear panel data model can be represented as follows (in notation based on Arellano
(2003)):
yit = αyi,t−1 + β0 xit + ηi + vit (21.1)
The main idea behind the difference estimator is to sweep out the individual effect via differencing.
First-differencing eq. (21.1) yields
in obvious notation. The error term of (21.2) is, by construction, autocorrelated and also correlated
with the lagged dependent variable, so an estimator that takes both issues into account is needed.
The endogeneity issue is solved by noting that all values of yi,t−k with k > 1 can be used as
instruments for ∆yi,t−1 : unobserved values of yi,t−k (whether missing or pre-sample) can safely be
substituted with 0. In the language of GMM, this amounts to using the relation
as an orthogonality condition.
Autocorrelation is dealt with by noting that if vit is white noise, the covariance matrix of the vector
whose typical element is ∆vit is proportional to a matrix H that has 2 on the main diagonal, −1
on the first subdiagonals and 0 elsewhere. One-step GMM estimation of equation (21.2) amounts to
computing
−1
N N N N
0 0 0 0
X X X X
γ̂ = W iZ i AN Z i W i W iZ i AN y i
Z i ∆y (21.4)
i=1 i=1 i=1 i=1
where
h i0
yi =
∆y ∆yi,3 ··· ∆yi,T
" #0
∆yi,2 ··· ∆yi,T −1
Wi =
∆xi,3 ··· ∆xi,T
0
yi1 0 0 ··· 0 ∆xi3
0 yi1 yi2 ··· 0
∆xi4
Zi =
..
.
0 0 0 ··· yi,T −2 ∆xiT
186
Chapter 21. Dynamic panel models 187
and
−1
N
Z 0i HZ
X
AN = Z i
i=1
Once the 1-step estimator is computed, the sample covariance matrix of the estimated residuals
can be used instead of H to obtain 2-step estimates, which are not only consistent but asymp-
totically efficient. (In principle the process may be iterated, but nobody seems to be interested.)
Standard GMM theory applies, except for one thing: Windmeijer (2005) has computed finite-sample
corrections to the asymptotic covariance matrix of the parameters, which are nowadays almost
universally used.
The difference estimator is consistent, but has been shown to have poor properties in finite samples
when α is near one. People these days prefer the so-called “system” estimator, which complements
the differenced data (with lagged levels used as instruments) with data in levels (using lagged dif-
ferences as instruments). The system estimator relies on an extra orthogonality condition which
has to do with the earliest value of the dependent variable yi,1 . The interested reader is referred
to Blundell and Bond (1998, pp. 124–125) for details, but here it suffices to say that this condi-
tion is satisfied in mean-stationary models and brings an improvement in efficiency that may be
substantial in many cases.
The set of orthogonality conditions exploited in the system approach is not very much larger than
with the difference estimator since most of the possible orthogonality conditions associated with
the equations in levels are redundant, given those already used for the equations in differences.
The key equations of the system estimator can be written as
−1
N N N N
X 0 X 0 X 0 X 0
γ̃ = W Z̃
W̃ Z AN Z W̃
Z̃ W W Z̃
W̃ Z AN Z ∆ỹ
Z̃ y i (21.5)
i=1 i=1 i=1 i=1
where
h i0
yi =
∆ỹ ∆yi3 ··· ∆yiT yi3 ··· yiT
" #0
∆yi2 ··· ∆yi,T −1 yi2 ··· yi,T −1
Wi =
W̃
∆xi3 ··· ∆xiT xi3 ··· xiT
0
yi1 0 0 ··· 0 0 ··· 0 ∆xi,3
0 yi1 yi2 ··· 0 0 ··· 0 ∆xi,4
..
.
0 0 0 ··· yi,T −2 0 ··· 0 ∆xiT
Zi =
Z̃ ..
.
0 0 0 ··· 0 ··· 0 xi3
∆yi2
..
.
0 0 0 ··· 0 0 ··· ∆yi,T −1 xiT
and
−1
N
X 0
AN = Z H ∗ Z̃
Z̃ Z
i=1
In this case choosing a precise form for the matrix H ∗ for the first step is no trivial matter. Its
north-west block should be as similar as possible to the covariance matrix of the vector ∆vit , so
Chapter 21. Dynamic panel models 188
the same choice as the “difference” estimator is appropriate. Ideally, the south-east block should
be proportional to the covariance matrix of the vector ι ηi + v , that is σv2 I + ση2 ι ι 0 ; but since ση2 is
unknown and any positive definite matrix renders the estimator consistent, people just use I. The
off-diagonal blocks should, in principle, contain the covariances between ∆vis and vit , which would
be an identity matrix if vit is white noise. However, since the south-east block is typically given a
conventional value anyway, the benefit in making this choice is not obvious. Some packages use I;
others use a zero matrix. Asymptotically, it should not matter, but on real datasets the difference
between the resulting estimates can be noticeable.
Rank deficiency
Both the difference estimator (21.4) and the system estimator (21.5) depend for their existence on
the invertibility of A N . This matrix may turn out to be singular for several reasons. However, this
does not mean that the estimator is not computable: in some cases, adjustments are possible such
that the estimator does exist, but the user should be aware that in these cases not all software
packages use the same strategy and replication of results may prove difficult or even impossible.
A first reason why A N may be singular could be the unavailability of instruments, chiefly because
of missing observations. This case is easy to handle. If a particular row of Z̃ Z i is zero for all
units, the corresponding orthogonality condition (or the corresponding instrument if you prefer) is
automatically dropped; of course, the overidentification rank is adjusted for testing purposes.
Even if no instruments are zero, however, A N could be rank deficient. A trivial case occurs if there
are collinear instruments, but a less trivial case may arise when T (the total number of time periods
available) is not much smaller than N (the number of units), as, for example, in some macro datasets
where the units are countries. The total number of potentially usable orthogonality conditions is
O(T 2 ), which may well exceed N in some cases. Of course A N is the sum of N matrices which have,
at most, rank 2T − 3 and therefore it could well happen that the sum is singular.
In all these cases, what we consider the “proper” way to go is to substitute the pseudo-inverse of
A N (Moore–Penrose) for its regular inverse. Again, our choice is shared by some software packages,
but not all, so replication may be hard.
yt = αyt−1 + η + t
where the i index is omitted for clarity. Suppose you have an individual with t = 1 . . . 5, for which
y3 is missing. It may seem that the data for this individual are unusable, because differencing yt
would produce something like
t 1 2 3 4 5
yt ∗ ∗ ◦ ∗ ∗
∆yt ◦ ∗ ◦ ◦ ∗
where ∗ = nonmissing and ◦ = missing. Estimation seems to be unfeasible, since there are no
periods in which ∆yt and ∆yt−1 are both observable.
However, we can use a k-difference operator and get
∆k yt = α∆k yt−1 + ∆k t
where ∆k = 1 − Lk and past levels of yt are perfectly valid instruments. In this example, we can
choose k = 3 and use y1 as an instrument, so this unit is in fact perfectly usable.
Not all software packages seem to be aware of this possibility, so replicating published results may
prove tricky if your dataset contains individuals with gaps between valid observations.
Chapter 21. Dynamic panel models 189
21.2 Usage
One of the concepts underlying the syntax of dpanel is that you get default values for several
choices you may want to make, so that in a “standard” situation the command is very concise. The
simplest case of the model (21.1) is a plain AR(1) process:
dpanel 1 ; y
gretl assumes that you want to estimate (21.6) via the difference estimator (21.4), using as many
orthogonality conditions as possible. The scalar 1 between dpanel and the semicolon indicates that
only one lag of y is included as an explanatory variable; using 2 would give an AR(2) model. The
syntax that gretl uses for the non-seasonal AR and MA lags in an ARMA model is also supported in
this context.1 For example, if you want the first and third lags of y (but not the second) included as
explanatory variables you can say
dpanel {1 3} ; y
To use a single lag of y other than the first you need to employ this mechanism:
To use the system estimator instead, you add the --system option, as in
dpanel 1 ; y --system
The level orthogonality conditions and the corresponding instrument are appended automatically
(see eq. 21.5).
Regressors
If we want to introduce additional regressors, we list them after the dependent variable in the same
way as other gretl commands, such as ols.
For the difference orthogonality relations, dpanel takes care of transforming the regressors in par-
allel with the dependent variable. Note that this differs from gretl’s arbond command, where only
the dependent variable is differenced automatically; it brings us more in line with other software.
One case of potential ambiguity is when an intercept is specified but the difference-only estimator
is selected, as in
dpanel 1 ; y const
In this case the default dpanel behavior, which agrees with Stata’s xtabond2, is to drop the con-
stant (since differencing reduces it to nothing but zeros). However, for compatibility with the
DPD package for Ox, you can give the option --dpdstyle, in which case the constant is retained
1 This represents an enhancement over the arbond command.
Chapter 21. Dynamic panel models 190
(equivalent to including a linear trend in equation 21.1). A similar point applies to the period-
specific dummy variables which can be added in dpanel via the --time-dummies option: in the
differences-only case these dummies are entered in differenced form by default, but when the
--dpdstyle switch is applied they are entered in levels.
The standard gretl syntax applies if you want to use lagged explanatory variables, so for example
the command
Instruments
The default rules for instruments are:
• lags of the dependent variable are instrumented using all available orthogonality conditions;
and
• additional regressors are considered exogenous, so they are used as their own instruments.
If a different policy is wanted, the instruments should be specified in an additional list, separated
from the regressors list by a semicolon. The syntax closely mirrors that for the tsls command,
but in this context it is necessary to distinguish between “regular” instruments and what are often
called “GMM-style” instruments (that is, instruments that are handled in the same block-diagonal
manner as lags of the dependent variable, as described above).
“Regular” instruments are transformed in the same way as regressors, and the contemporaneous
value of the transformed variable is used to form an orthogonality condition. Since regressors are
treated as exogenous by default, it follows that these two commands estimate the same model:
dpanel 1 ; y z
dpanel 1 ; y z ; z
The instrument specification in the second case simply confirms what is implicit in the first: that
z is exogenous. Note, though, that if you have some additional variable z2 which you want to add
as a regular instrument, it then becomes necessary to include z in the instrument list if it is to be
treated as exogenous:
The specification of “GMM-style” instruments is handled by the special constructs GMM() and
GMMlevel(). The first of these relates to instruments for the equations in differences, and the
second to the equations in levels. The syntax for GMM() is
where name is replaced by the name of a series (or the name of a list of series), and minlag and
maxlag are replaced by the minimum and maximum lags to be used as instruments. The same goes
for GMMlevel().
One common use of GMM() is to limit the number of lagged levels of the dependent variable used
as instruments for the equations in differences. It’s well known that although exploiting all pos-
sible orthogonality conditions yields maximal asymptotic efficiency, in finite samples it may be
preferable to use a smaller subset (but see also Okui (2009)). For example, the specification
Chapter 21. Dynamic panel models 191
dpanel 1 ; y ; GMM(y, 2, 4)
dpanel 1 ; y x ; GMM(z, 2, 6)
the variable x is considered an endogenous regressor, and up to 5 lags of z are used as instruments.
Note that in the following script fragment
dpanel 1 ; y z
dpanel 1 ; y z ; GMM(z,0,0)
the two estimation commands should not be expected to give the same result, as the sets of orthog-
onality relationships are subtly different. In the latter case, you have T − 2 separate orthogonality
relationships pertaining to zit , none of which has any implication for the other ones; in the former
case, you only have one. In terms of the Z i matrix, the first form adds a single row to the bottom
of the instruments matrix, while the second form adds a diagonal block with T − 2 columns; that
is, h i
zi3 zi4 · · · zit
versus
zi3 0 ··· 0
0 zi4 ··· 0
.. ..
. .
0 0 ··· zit
#include <oxstd.h>
#import <packages/dpd/dpd>
2 Seehttp://www.doornik.com/download.html.
3 Tobe specific, this is using Ox Console version 5.10, version 1.24 of the DPD package, and gretl built from CVS as of
2010-10-23, all on Linux.
Chapter 21. Dynamic panel models 192
main()
{
decl dpd = new DPD();
dpd.Load("abdata.in7");
dpd.SetYear("YEAR");
delete dpd;
}
In the examples below we take this template for granted and show just the model-specific code.
Example 1
The following Ox/DPD code—drawn from abest1.ox— replicates column (b) of Table 4 in Arellano
and Bond (1991), an instance of the differences-only or GMM-DIF estimator. The dependent variable
is the log of employment, n; the regressors include two lags of the dependent variable, current and
lagged values of the log real-product wage, w, the current value of the log of gross capital, k, and
current and lagged values of the log of industry output, ys. In addition the specification includes
a constant and five year dummies; unlike the stochastic regressors, these deterministic terms are
not differenced. In this specification the regressors w, k and ys are treated as exogenous and serve
as their own instruments. In DPD syntax this requires entering these variables twice, on the X_VAR
and I_VAR lines. The GMM-type (block-diagonal) instruments in this example are the second and
subsequent lags of the level of n. Both 1-step and 2-step estimates are computed.
dpd.Gmm("n", 2, 99);
dpd.SetDummies(D_CONSTANT + D_TIME);
open abdata.gdt
list X = w w(-1) k ys ys(-1)
dpanel 2 ; n X const --time-dummies --asy --dpdstyle
dpanel 2 ; n X const --time-dummies --asy --two-step --dpdstyle
Note that in gretl the switch to suppress robust standard errors is --asymptotic, here abbreviated
to --asy.4 The --dpdstyle flag specifies that the constant and dummies should not be differenced,
in the context of a GMM-DIF model. With gretl’s dpanel command it is not necessary to specify the
exogenous regressors as their own instruments since this is the default; similarly, the use of the
second and all longer lags of the dependent variable as GMM-type instruments is the default and
need not be stated explicitly.
4 Option flags in gretl can always be truncated, down to the minimal unique abbreviation.
Chapter 21. Dynamic panel models 193
Example 2
The DPD file abest3.ox contains a variant of the above that differs with regard to the choice of
instruments: the variables w and k are now treated as predetermined, and are instrumented GMM-
style using the second and third lags of their levels. This approximates column (c) of Table 4 in
Arellano and Bond (1991). We have modified the code in abest3.ox slightly to allow the use of
robust (Windmeijer-corrected) standard errors, which are the default in both DPD and gretl with
2-step estimation:
dpd.Gmm("n", 2, 99);
dpd.Gmm("w", 2, 3);
dpd.Gmm("k", 2, 3);
open abdata.gdt
list X = w w(-1) k ys ys(-1)
list Ivars = ys ys(-1)
dpanel 2 ; n X const ; GMM(w,2,3) GMM(k,2,3) Ivars --time --two-step --dpd
Note that since we are now calling for an instrument set other then the default (following the second
semicolon), it is necessary to include the Ivars specification for the variable ys. However, it is
not necessary to specify GMM(n,2,99) since this remains the default treatment of the dependent
variable.
Example 3
Our third example replicates the DPD output from bbest1.ox: this uses the same dataset as the
previous examples but the model specifications are based on Blundell and Bond (1998), and involve
comparison of the GMM-DIF and GMM-SYS (“system”) estimators. The basic specification is slightly
simplified in that the variable ys is not used and only one lag of the dependent variable appears as
a regressor. The Ox/DPD code is:
dpd.GmmLevel("k", 1, 1);
dpd.SetMethod(M_2STEP);
dpd.Estimate();
open abdata.gdt
list X = w w(-1) k k(-1)
list Z = w k
Note the use of the --system option flag to specify GMM-SYS, including the default treatment of
the dependent variable, which corresponds to GMMlevel(n,1,1). In this case we also want to
use lagged differences of the regressors w and k as instruments for the levels equations so we
need explicit GMMlevel entries for those variables. If you want something other than the default
treatment for the dependent variable as an instrument for the levels equations, you should give an
explicit GMMlevel specification for that variable — and in that case the --system flag is redundant
(but harmless).
For the sake of completeness, note that if you specify at least one GMMlevel term, dpanel will then
include equations in levels, but it will not automatically add a default GMMlevel specification for
the dependent variable unless the --system option is given.
which allows for a time-specific disturbance νt . The Solow model with Cobb–Douglas production
function implies that γ = −α, but this assumption is not imposed in estimation. The time-specific
disturbance is eliminated by subtracting the period mean from each of the series.
5 We say an “approximation” because we have not been able to replicate exactly the OLS results reported in the papers
cited, though it seems from the description of the data in Caselli et al. (1996) that we ought to be able to do so. We note
that Bond et al. (2001) used data provided by Professor Caselli yet did not manage to reproduce the latter’s results.
Chapter 21. Dynamic panel models 195
Equation (21.7) can be transformed to an AR(1) dynamic panel-data model by adding yi,t−5 to both
sides, which gives
yit = (1 + β)yi,t−5 + αsit + γ(nit + g + δ) + ηi + it (21.8)
where all variables are now assumed to be time-demeaned.
In (rough) replication of Bond et al. (2001) we now proceed to estimate the following two models:
(a) equation (21.8) via GMM-DIF, using as instruments the second and all longer lags of yit , sit and
nit + g + δ; and (b) equation (21.8) via GMM-SYS, using ∆yi,t−1 , ∆si,t−1 and ∆(ni,t−1 + g + δ) as
additional instruments in the levels equations. We report robust standard errors throughout. (As a
purely notational matter, we now use “t − 1” to refer to values five years prior to t, as in Bond et al.
(2001)).
The gretl script to do this job is shown below. Note that the final transformed versions of the
variables (logs, with time-means subtracted) are named ly (yit ), linv (sit ) and lngd (nit + g + δ).
open CEL.gdt
ngd = n + 0.05
ly = log(y)
linv = log(s)
lngd = log(ngd)
smpl --full
list X = linv lngd
# 1-step GMM-DIF
dpanel 1 ; ly X ; GMM(X,2,99)
# 2-step GMM-DIF
dpanel 1 ; ly X ; GMM(X,2,99) --two-step
# GMM-SYS
dpanel 1 ; ly X ; GMM(X,2,99) GMMlevel(X,1,1) --two-step --sys
For comparison we estimated the same two models using Ox/DPD and the Stata command xtabond2.
(In each case we constructed a comma-separated values dataset containing the data as transformed
in the gretl script shown above, using a missing-value code appropriate to the target program.) For
reference, the commands used with Stata are reproduced below:
For the GMM-DIF model all three programs find 382 usable observations and 30 instruments, and
yield identical parameter estimates and robust standard errors (up to the number of digits printed,
or more); see Table 21.1.6
6 The coefficient shown for ly(-1) in the Tables is that reported directly by the software; for comparability with the
original model (eq. 21.7) it is necesary to subtract 1, which produces the expected negative value indicating conditional
convergence in per capita income.
Chapter 21. Dynamic panel models 196
1-step 2-step
coeff std. error coeff std. error
ly(-1) 0.577564 0.1292 0.610056 0.1562
linv 0.0565469 0.07082 0.100952 0.07772
lngd −0.143950 0.2753 −0.310041 0.2980
Results for GMM-SYS estimation are shown in Table 21.2. In this case we show two sets of gretl
results: those labeled “gretl(1)” were obtained using gretl’s --dpdstyle option, while those labeled
“gretl(2)” did not use that option—the intent being to reproduce the H matrices used by Ox/DPD
and xtabond2 respectively.
In this case all three programs use 479 observations; gretl and xtabond2 use 41 instruments and
produce the same estimates (when using the same H matrix) while Ox/DPD nominally uses 66.7
It is noteworthy that with GMM-SYS plus “messy” missing observations, the results depend on the
precise array of instruments used, which in turn depends on the details of the implementation of
the estimator.
v∗
where the v̂ i are the transformed (e.g. differenced) residuals for unit i. Under the null hypothesis
that the instruments are valid, S is asymptotically distributed as chi-square with degrees of freedom
equal to the number of overidentifying restrictions.
In general we see a good level of agreement between gretl, DPD and xtabond2 with regard to these
statistics, with a few relatively minor exceptions. Specifically, xtabond2 computes both a “Sargan
test” and a “Hansen test” for overidentification, but what it calls the Hansen test is, apparently, what
DPD calls the Sargan test. (We have had difficulty determining from the xtabond2 documentation
(Roodman, 2006) exactly how its Sargan test is computed.) In addition there are cases where the
degrees of freedom for the Sargan test differ between DPD and gretl; this occurs when the A N
matrix is singular (section 21.1). In concept the df equals the number of instruments minus the
7 This is a case of the issue described in section 21.1: the full A matrix turns out to be singular and special measures
N
must be taken to produce estimates.
Chapter 21. Dynamic panel models 197
number of parameters estimated; for the first of these terms gretl uses the rank of A N , while DPD
appears to use the full dimension of this matrix.
Key Content
AR1, AR2 1st and 2nd order autocorrelation test statistics
sargan, sargan_df Sargan test for overidentifying restrictions and correspond-
ing degrees of freedom
wald, wald_df Wald test for overall significance and corresponding de-
grees of freedom
GMMinst The matrix Z of instruments (see equations (21.2) and (21.5)
wgtmat The matrix A of GMM weights (see equations (21.2) and
(21.5)
Note, however, that GMMinst and wgtmat (which may be quite large matrices) are not saved in
the $model bundle by default; that requires use of the --keep-extra option with the dpanel
command. The script in Table 21.6 illustrates use of these matrices to replicate via hansl commands
the calculation of the GMM estimator.
print coef
Note the command words nls (which introduces the regression function), deriv (which introduces
the specification of a derivative), and end nls, which terminates the specification and calls for
estimation. If the --vcv flag is appended to the last line the covariance matrix of the parameter
estimates is printed.
ols C 0 Y
genr alpha = $coeff(0)
genr beta = $coeff(Y)
genr gamma = 1
199
Chapter 22. Nonlinear least squares 200
ols y 0 x1 x2
genr alpha = $coeff(0)
genr beta = $coeff(x1)
# Greene
nls C = alpha + beta * Y^gamma
params alpha beta gamma
end nls
# Davidson
nls y = alpha + beta * x1 + (1/beta) * x2
params alpha beta
end nls
If analytical derivatives are supplied, they are checked for consistency with the given nonlinear
function. If the derivatives are clearly incorrect estimation is aborted with an error message. If the
Chapter 22. Nonlinear least squares 201
derivatives are “suspicious” a warning message is issued but estimation proceeds. This warning
may sometimes be triggered by incorrect derivatives, but it may also be triggered by a high degree
of collinearity among the derivatives.
Note that you cannot mix analytical and numerical derivatives: you should supply expressions for
all of the derivatives or none.
open greene25_1.gdt
list X = const age income ownrent selfempl
# initalisation
ols cardhldr X --quiet
matrix b = $coeff / $sigma
The example in script 22.1 can be enhanced by using analytical derivatives: since
xi)
∂g(x
b 0 x i ) · xij
= ϕ(b
∂bj
one could substitute the params line in the script with the two-liner
series f = dnorm(ndx)
deriv b = {f} .* {X}
and have nls use analytically-computed derivatives, which are quicker and usually more reliable.
Chapter 22. Nonlinear least squares 202
• Both the actual and predicted relative reductions in the error sum of squares are at most .
This default value of is the machine precision to the power 3/4,1 but it can be adjusted using the
set command with the parameter nls_toler. For example
1 On a 32-bit Intel Pentium machine a likely value for this parameter is 1.82 × 10−12 .
2 For a discussion of gretl’s accuracy in the estimation of linear models, see Appendix D.
Chapter 22. Nonlinear least squares 203
using all analytical derivatives and one using all numerical approximations. In each case the default
tolerance was used.3
Out of the 54 runs, gretl failed to produce a solution in 4 cases when using analytical derivatives,
and in 5 cases when using numeric approximation. Of the four failures in analytical derivatives
mode, two were due to non-convergence of the Levenberg–Marquardt algorithm after the maximum
number of iterations (on MGH09 and Bennett5, both described by NIST as of “Higher difficulty”) and
two were due to generation of range errors (out-of-bounds floating point values) when computing
the Jacobian (on BoxBOD and MGH17, described as of “Higher difficulty” and “Average difficulty”
respectively). The additional failure in numerical approximation mode was on MGH10 (“Higher diffi-
culty”, maximum number of iterations reached).
The table gives information on several aspects of the tests: the number of outright failures, the
average number of iterations taken to produce a solution and two sorts of measure of the accuracy
of the estimates for both the parameters and the standard errors of the parameters.
For each of the 54 runs in each mode, if the run produced a solution the parameter estimates
obtained by gretl were compared with the NIST certified values. We define the “minimum correct
figures” for a given run as the number of significant figures to which the least accurate gretl esti-
mate agreed with the certified value, for that run. The table shows both the average and the worst
case value of this variable across all the runs that produced a solution. The same information is
shown for the estimated standard errors.4
The second measure of accuracy shown is the percentage of cases, taking into account all parame-
ters from all successful runs, in which the gretl estimate agreed with the certified value to at least
the 6 significant figures which are printed by default in the gretl regression output.
Using analytical derivatives, the worst case values for both parameters and standard errors were
improved to 6 correct figures on the test machine when the tolerance was tightened to 1.0e−14.
3 The data shown in the table were gathered from a pre-release build of gretl version 1.0.9, compiled with gcc 3.3,
linked against glibc 2.3.2, and run under Linux on an i686 PC (IBM ThinkPad A21m).
4 For the standard errors, I excluded one outlier from the statistics shown in the table, namely Lanczos1. This is an
odd case, using generated data with an almost-exact fit: the standard errors are 9 or 10 orders of magnitude smaller
than the coefficients. In this instance gretl could reproduce the certified standard errors to only 3 figures (analytical
derivatives) and 2 figures (numerical derivatives).
Chapter 22. Nonlinear least squares 204
Using numerical derivatives, the same tightening of the tolerance raised the worst values to 5
correct figures for the parameters and 3 figures for standard errors, at a cost of one additional
failure of convergence.
Note the overall superiority of analytical derivatives: on average solutions to the test problems
were obtained with substantially fewer iterations and the results were more accurate (most notably
for the estimated standard errors). Note also that the six-digit results printed by gretl are not 100
percent reliable for difficult nonlinear problems (in particular when using numerical derivatives).
Having registered this caveat, the percentage of cases where the results were good to six digits or
better seems high enough to justify their printing in this form.
Chapter 23
T
X
`(θ) = `t (θ)
t=1
which is true in most cases of interest. The functions `t (θ) are called the log-likelihood contribu-
tions.
Moreover, the location of the maximum is obviously determined by the data Y. This means that the
value
θ̂(Y) = Argmax `(θ) (23.1)
θ∈Θ
is some function of the observed data (a statistic), which has the property, under mild conditions,
of being a consistent, asymptotically normal and asymptotically efficient estimator of θ.
Sometimes it is possible to write down explicitly the function θ̂(Y); in general, it need not be so. In
these circumstances, the maximum can be found by means of numerical techniques. These often
rely on the fact that the log-likelihood is a smooth function of θ, and therefore on the maximum
its partial derivatives should all be 0. The gradient vector, or score vector, is a function that enjoys
many interesting statistical properties in its own right; it will be denoted here as g(θ). It is a
1 We are supposing here that our data are a realization of continuous random variables. For discrete random variables,
everything continues to apply by referring to the probability function instead of the density. In both cases, the distribution
may be conditional on some exogenous variables.
205
Chapter 23. Maximum likelihood estimation 206
1. pick a point θ0 ∈ Θ;
2. evaluate g(θ0 );
5. substitute θ0 with θ1 ;
6. restart from 2.
Many algorithms of this kind exist; they basically differ from one another in the way they compute
the direction vector d(g(θ0 )), to ensure that `(θ1 ) > `(θ0 ) (so that we eventually end up on the
maximum).
The default method gretl uses to maximize the log-likelihood is a gradient-based algorithm known
as the BFGS (Broyden, Fletcher, Goldfarb and Shanno) method. This technique is used in most
econometric and statistical packages, as it is well-established and remarkably powerful. Clearly,
in order to make this technique operational, it must be possible to compute the vector g(θ) for
any value of θ. In some cases this vector can be written explicitly as a function of Y. If this is
not possible or too difficult the gradient may be evaluated numerically. The alternative Newton-
Raphson algorithm is also available. This method is more effective under some circumstances but
is also more fragile; see section 23.8 and chapter 34 for details.2
The choice of the starting value, θ0 , is crucial in some contexts and inconsequential in others. In
general, however, it is advisable to start the algorithm from “sensible” values whenever possible. If
a consistent estimator is available, this is usually a safe and efficient choice: this ensures that in
large samples the starting point will be likely close to θ̂ and convergence can be achieved in few
iterations.
The maximum number of iterations allowed for the BFGS procedure, and the relative tolerance
for assessing convergence, can be adjusted using the set command: the relevant variables are
bfgs_maxiter (default value 500) and bfgs_toler (default value, the machine precision to the
power 3/4).
where G(θ̂) is the T × k matrix of contributions to the gradient. Two other options are available. If
the --hessian flag is given, the covariance matrix is computed from a numerical approximation to
the Hessian at convergence. If the --robust option is selected, the quasi-ML “sandwich” estimator
is used:
d QML (θ̂) = H(θ̂)−1 G0 (θ̂)G(θ̂)H(θ̂)−1
Var
where H denotes the numerical approximation to the Hessian.
2 Note that some of the statements made below (for example, regarding estimation of the covariance matrix) have to
Cluster-robust estimation is also available: in order to activate it, use the --cluster=clustvar,
where clustvar should be a discrete series. See section 19.5 for more details.
Note, however, that if the log-likelihood function supplied by the user just returns a scalar value —
as opposed to a series or vector holding per-observation contributions — then the OPG method is
not applicable and so the covariance matrix must be estimated via a numerical approximation to
the Hessian.
αp p−1
f (xt ) = x exp (−αxt ) (23.3)
Γ (p) t
The log-likelihood for the entire sample can be written as the logarithm of the joint density of all
the observations. Since these are independent and identical, the joint density is the product of the
individual densities, and hence its log is
T T
" #
X αp p−1 X
`(α, p) = log xt exp (−αxt ) = `t (23.4)
t=1
Γ (p) t=1
where
`t = p · log(αxt ) − γ(p) − log xt − αxt
and γ(·) is the log of the gamma function. In order to estimate the parameters α and p via ML, we
need to maximize (23.4) with respect to them. The corresponding gretl code snippet is
scalar alpha = 1
scalar p = 1
alpha = 1
p = 1
are necessary to ensure that the variables alpha and p exist before the computation of logl is
attempted. Inside the mle block these variables (which could be either scalars, vectors or a com-
bination of the two — see below for an example) are identified as the parameters that should be
adjusted to maximize the likelihood via the params keyword. Their values will be changed by the
execution of the mle command; upon successful completion, they will be replaced by the ML esti-
mates. The starting value is 1 for both; this is arbitrary and does not matter much in this example
(more on this later).
The above code can be made more readable, and marginally more efficient, by defining a variable
to hold α · xt . This command can be embedded in the mle block as follows:
The variable ax is not added to the params list, of course, since it is just an auxiliary variable to
facilitate the calculations. You can insert as many such auxiliary lines as you require before the
params line, with the restriction that they must contain either (a) commands to generate series,
scalars or matrices or (b) print commands (which may be used to aid in debugging).
In a simple example like this, the choice of the starting values is almost inconsequential; the algo-
rithm is likely to converge no matter what the starting values are. However, consistent method-of-
moments estimators of p and α can be simply recovered from the sample mean m and variance V :
since it can be shown that
E(xt ) = p/α V (xt ) = p/α2
it follows that the following estimators
ᾱ = m/V
p̄ = m · ᾱ
are consistent, and therefore suitable to be used as starting point for the algorithm. The gretl script
code then becomes
scalar m = mean(x)
scalar alpha = m/var(x)
scalar p = m*alpha
Another thing to note is that sometimes parameters are constrained within certain boundaries: in
this case, for example, both α and p must be positive numbers. Gretl does not check for this: it
is the user’s responsibility to ensure that the function is always evaluated at an admissible point
in the parameter space during the iterative search for the maximum. An effective technique is to
define a variable for checking that the parameters are admissible and setting the log-likelihood as
undefined if the check fails. An example, which uses the conditional assignment operator, follows:
scalar m = mean(x)
scalar alpha = m/var(x)
scalar p = m*alpha
linear specification for Ci∗ is often chosen. For example, the Cobb–Douglas cost function arises
when Ci∗ is a linear function of the logarithms of the input prices and the output quantities.
The stochastic frontier model is a linear model of the form yi = xi β + εi in which the error term εi
is the sum of ui and vi .
A common postulate is that ui ∼ N(0, σ 2 ) and vi ∼ N(0, σ 2 ). If independence between ui and
u v
vi is also assumed, then it is possible to show that the density function of εi has the form:
s
2 λεi 1 εi
f (εi ) = φ (23.5)
π σ σ σ
Φ
q Φ(·) and φ(·) are, respectively, the distribution and density function of the standard normal,
where
σ = σu2 + σv2 and λ = σ u
σv .
As a consequence, the log-likelihood for one observation takes the form (apart form an irrelevant
constant)
εi2
" #
λεi
`t = log Φ − log(σ ) +
σ 2σ 2
Therefore, a Cobb–Douglas cost function with stochastic frontier is the model described by the
following equations:
In most cases, one wants to ensure that the homogeneity of the cost function
Pn with respect to
the prices holds by construction. Since this requirement is equivalent to j=1 αj = 1, the above
equation for Ci∗ can be rewritten as
m
X n
X
log Ci − log pin = c + βj log yij + αj (log pij − log pin ) + εi (23.6)
j=1 j=2
The above equation could be estimated by OLS, but it would suffer from two drawbacks: first,
the OLS estimator for the intercept c is inconsistent because the disturbance term has a non-zero
expected value; second, the OLS estimators for the other parameters are consistent, but inefficient
in view of the non-normality of εi . Both issues can be addressed by estimating (23.6) by maximum
likelihood. Nevertheless, OLS estimation is a quick and convenient way to provide starting values
for the MLE algorithm.
Listing 23.1 shows how to implement the model described so far. The banks91 file contains part
of the data used in Lucchetti, Papi and Zazzaro (2001).
The script in example 23.1 is relatively easy to modify to show how one can use vectors (that is,
1-dimensional matrices) for storing the parameters to optimize: example 23.2 holds essentially the
same script in which the parameters of the cost function are stored together in a vector. Of course,
this makes also possible to use variable lists and other refinements which make the code more
compact and readable.
Chapter 23. Maximum likelihood estimation 210
Listing 23.1: Estimation of stochastic frontier cost function (with scalar parameters)
open banks91.gdt
# transformations
series cost = ln(VC)
series q1 = ln(Q1)
series q2 = ln(Q2)
series p1 = ln(P1)
series p2 = ln(P2)
series p3 = ln(P3)
scalar b0 = $coeff(const)
scalar b1 = $coeff(q1)
scalar b2 = $coeff(q2)
scalar b3 = $coeff(rp2)
scalar b4 = $coeff(rp3)
scalar su = 0.1
scalar sv = 0.1
Listing 23.2: Estimation of stochastic frontier cost function (with matrix parameters)
open banks91.gdt
# transformations
series cost = ln(VC)
series q1 = ln(Q1)
series q2 = ln(Q2)
series p1 = ln(P1)
series p2 = ln(P2)
series p3 = ln(P3)
ols rcost X
X = const q1 q2 rp2 rp3
# Cobb-Douglas cost function with homogeneity restrictions
# and inefficiency
matrix b = $coeff
scalar su = 0.1
scalar sv = 0.1
yt = µ + εt
εt = ut · σt
ut ∼ N(0, 1)
2
ht = ω + αεt−1 + βht−1 .
Since the variance of yt depends on past values, writing down the log-likelihood function is not
simply a matter of summing the log densities for individual observations. As is common in time
series models, yt cannot be considered independent of the other observations in our sample, and
consequently the density function for the whole sample (the joint density for all observations) is
not just the product of the marginal densities.
Maximum likelihood estimation, in these cases, is achieved by considering conditional densities, so
what we maximize is a conditional likelihood function. If we define the information set at time t as
Ft = yt , yt−1 , . . . ,
yt |Ft−1 ∼ N [µ, ht ] .
By means of the properties of conditional distributions, the joint density can be factorized as
follows
T
Y
f (yt , yt−1 , . . .) = f (yt |Ft−1 ) · f (y0 )
t=1
If we treat y0 as fixed, then the term f (y0 ) does not depend on the unknown parameters, and there-
fore the conditional log-likelihood can then be written as the sum of the individual contributions
as
XT
`(µ, ω, α, β) = `t (23.7)
t=1
The following script shows a simple application of this technique, which uses the data file djclose;
it is one of the example dataset supplied with gretl and contains daily data from the Dow Jones
stock index.
open djclose
series y = 100*ldiff(djclose)
scalar mu = 0.0
scalar omega = 1
scalar alpha = 0.4
scalar beta = 0.0
series e = y - mu
series h = var(y)
series h = omega + alpha*(e(-1))^2 + beta*h(-1)
params mu omega alpha beta
end mle
set bfgs_richardson on
However, one might want to avoid the approximation and specify an exact function for the deriva-
tives. As an example, consider the following script:
nulldata 1000
genr x1 = normal()
genr x2 = normal()
genr x3 = normal()
scalar b0 = 0
scalar b1 = 0
scalar b2 = 0
scalar b3 = 0
Here, 1000 data points are artificially generated for an ordinary probit model:4 yt is a binary
variable, which takes the value 1 if yt∗ = β1 x1t + β2 x2t + β3 x3t + εt > 0 and 0 otherwise. Therefore,
yt = 1 with probability Φ(β1 x1t + β2 x2t + β3 x3t ) = πt . The probability function for one observation
can be written as
y
P (yt ) = πt t (1 − πt )1−yt
4 Again, gretl does provide a native probit command (see section 35.1), but a probit model makes for a nice example
here.
Chapter 23. Maximum likelihood estimation 214
Since the observations are independent and identically distributed, the log-likelihood is simply the
sum of the individual contributions. Hence
T
X
`= yt log(πt ) + (1 − yt ) log(1 − πt )
t=1
The --verbose switch at the end of the end mle statement produces a detailed account of the
iterations done by the BFGS algorithm.
In this case, numerical differentiation works rather well; nevertheless, computation of the analytical
∂`
score is straightforward, since the derivative ∂βi can be written as
∂` ∂` ∂πt
= ·
∂βi ∂πt ∂βi
via the chain rule, and it is easy to see that
∂` yt 1 − yt
= −
∂πt πt 1 − πt
∂πt
= φ(β1 x1t + β2 x2t + β3 x3t ) · xit
∂βi
The mle block in the above script can therefore be modified as follows:
Note that the params statement has been replaced by a series of deriv statements; these have the
double function of identifying the parameters over which to optimize and providing an analytical
expression for their respective score elements.
For the purpose of debugging ML estimators one additional sort of statement is allowed: you can
print the value of a relevant variable at each step of the iteration. This facility is more restricted
then the regular print command. The command word print should be followed by the name of
just one variable (a scalar, series or matrix).
In the last example above a key variable named m was generated, forming the basis for the analytical
derivatives. To track the progress of this variable one could add a print statement within the ML
block, as in
mle ll = logprob
series xb = exp(b0 + b1 * x)
series d = (y=0)
series poiprob = exp(-xb) * xb^y / gamma(y+1)
series logprob = (alpha>0) && (alpha<1) ? \
log(alpha*d + (1-alpha)*poiprob) : NA
params alpha b0 b1
end mle -v
However, the code above has to be modified each time we change our specification by, say, adding
an explanatory variable. Using functions, we can simplify this task considerably and eventually be
able to write something easy like
list X = const x
zip(y, X)
Let’s see how this can be done. First we need to define a function called zip() that will take two ar-
guments: a dependent variable y and a list of explanatory variables X. An example of such function
can be seen in script 23.3. By inspecting the function code, you can see that the actual estimation
does not happen here: rather, the zip() function merely uses the built-in modprint command to
print out the results coming from another user-written function, namely zip_estimate().
The function zip_estimate() is not meant to be executed directly; it just contains the number-
crunching part of the job, whose results are then picked up by the end function zip(). In turn,
zip_estimate() calls other user-written functions to perform other tasks. The whole set of “in-
ternal” functions is shown in the panel 23.4.
All the functions shown in 23.3 and 23.4 can be stored in a separate inp file and executed once, at
the beginning of our job, by means of the include command. Assuming the name of this script file
5 The actual ZIP model is in fact a bit more general than the one presented here. The specialized version discussed in
this section was chosen for the sake of simplicity. For futher details, see Greene (2003).
Chapter 23. Maximum likelihood estimation 216
/*
user-level function: estimate the model and print out
the results
*/
function void zip(series y, list X)
matrix coef_stde = zip_estimate(y, X)
printf "\nZero-inflated Poisson model:\n"
string parnames = "alpha,"
string parnames += varname(X)
modprint coef_stde parnames
end function
is zip_est.inp, the following is an example script which (a) includes the script file, (b) generates a
simulated dataset, and (c) performs the estimation of a ZIP model on the artificial data.
A further step may then be creating a function package for accessing your new zip() function via
gretl’s graphical interface. For details on how to do this, see section 13.5.
In the rest of this section, we will expand on the probit example of section 23.5 to give the reader
an idea of what a “heavy-duty” application of mle looks like. Most of the code fragments come from
mle-advanced.inp, which is one of the sample scripts supplied with the standard installation of
gretl (see under File > Script files > Practice File).
list X = const x1 x2 x3
matrix b = zeros(nelem(X),1)
In this context, the fact that the model we are estimating has four explanatory variables is totally
incidental: the code is written in such a way that we could change the content of the list X without
having to make any other modification. This was made possible by:
1. gathering the parameters to estimate into a single vector b rather than using separate scalars;
2. using the nelem() function to initialize b, so that its dimension is kept track of automatically;
A parallel enhancement could be achieved in the case of analytically computed derivatives: since
b is now a vector, mle expects the argument to the deriv keyword to be a matrix, in which each
column is the partial derivative to the corresponding element of b. It is useful to re-write the score
for the i-th observation as
∂`i
= mi x0i (23.8)
∂β
where mi is the “signed Mills’ ratio”, that is
φ(x0i β) φ(x0i β)
mi = yi − (1 − yi ) ,
Φ(x0i β) 1 − Φ(x0i β)
series P = cnorm(ndx)
series m = dnorm(ndx)*(y/P - (1-y)/(1-P))
and make use of the conditional assignment operator and of the specialized function invmills()
for efficiency. Building the score matrix is now easily achieved via
in which the {} operator was used to turn series and lists into matrices (see chapter 16). However,
proceeding in this way for more complex models than probit may imply inserting into the mle
block a long series of instructions; the example above merely happens to be short because the
score matrix for the probit model is very easy to write in matrix form.
A better solution is writing a user-level function to compute the score and using that inside the mle
block, as in
Chapter 23. Maximum likelihood estimation 219
[...]
In this way, no matter how complex the computation of the score is, the mle block remains nicely
compact.
where H(θ) is the Hessian of the total loglikelihood computed at θ and 0 < λ < 1 is a scalar called
the step length.
The above expression makes a few points clear:
1. At each step, it must be possible to compute not only the score g(θ), but also its derivative
H(θ);
3. it is assumed that for some positive value of λ, `(θ1 ) > `(θ0 ); in other words, that going in
the direction d [g(θ0 )] leads upwards for some step length.
The strength of Newton’s method lies in the fact that, if the loglikelihood is globally concave,
then (23.9) enjoys certain optimality properties and the number of iterations required to reach the
maximum is often much smaller than it would be with other methods, such as BFGS. However, it
may have some disadvantages: for a start, the Hessian H(θ) may be difficult or very expensive to
compute; moreover, the loglikelihood may not be globally concave, so for some values of θ, the
matrix H(θ) is not negative definite or perhaps even singular. Those cases are handled by gretl’s
implementation of Newton’s algorithm by means of several heuristic techniques6 , but a number of
adverse consequences may occur, which range from longer computation time for optimization to
non-convergence of the algorithm.
As a consequence, using Newton’s method is advisable only when the computation of the Hessian
is not too CPU-intensive and the nature of the estimator is such that it is known in advance that
the loglikelihood is globally concave. The probit models satisfies both requisites, so we will expand
the preceding example to illustrate how to use Newton’s method in gretl.
A first example may be given simply by issuing the command
scalar; however, if you’re interested in the precise details, you’ll be much better off looking at the source code: the file
you’ll want to look at is lib/src/gretl_bfgs.c.
Chapter 23. Maximum likelihood estimation 220
before the mle block.7 This will instruct gretl to use Newton’s method instead of BFGS. If the deriv
keyword is used, gretl will differentiate the score function numerically; otherwise, if the score
has to be computed itself numerically, gretl will calculate H(θ) by differentiating the loglikelihood
numerically twice. The latter solution, though, is generally to be avoided, as may be extremely
time-consuming and may yield imprecise results.
A much better option is to calculate the Hessian analytically and have gretl use its true value rather
than a numerical approximation. In most cases, this is both much faster and numerically stable,
but of course comes at the price of having to differentiate the loglikelihood twice to respect with
the parameters and translate the resulting expressions into efficient hansl code.
Luckily, both tasks are relatively easy in the probit case: the matrix of second derivatives of `i may
be written as
∂ 2 `i
0
= −m i mi + xi β xi x0i
∂β∂β0
so the total Hessian is
w1
n w2
X ∂ 2 `i
0
= −X X (23.10)
∂β∂β0
..
i=1
.
wn
where wi = mi mi + x0i β . It can be shown that wi > 0, so the Hessian is guaranteed to be negative
definite in all sensible cases and the conditions are ideal for applying Newton’s method.
A hansl translation of equation (23.10) may look like
There are two characteristics worth noting of the function above. For a start, it doesn’t return
anything: the result of the computation is simply stored in the matrix pointed at by the first
argument of the function. Second, the result is not the Hessian proper, but rather its negative. This
function becomes usable from within an mle block by the keyword hessian. The syntax is
mle ...
...
hessian funcname(&mat_addr, ...)
end mle
In other words, the hessian keyword must be followed by the call to a function whose first argu-
ment is a matrix pointer which is supposed to be filled with the negative of the Hessian at θ.
We said above (section 23.1) that the covariance matrix of the parameter estimates is by default
estimated using the Outer Product of the Gradient (so long as the log-likelihood function returns
the per-observation contributions). However, if you supply a function that computes the Hessian
then by default it is used in estimating the covariance matrix. If you wish to impose use of OPG
instead, append the --opg option to the end of the mle block.
Note that gretl does not perform any numerical check on whether a user-supplied function com-
putes the Hessian correctly. On the one hand, this means that you can trick mle into using alter-
natives to the Hessian and thereby implement other optimization methods. For example, if you
7 To go back to BFGS, you use set optimizer bfgs.
Chapter 23. Maximum likelihood estimation 221
substitute in equation 23.9 the Hessian H with the negative of the OPG matrix −G0 G, as defined in
(23.2), you get the so-called BHHH optimization method (see Berndt et al. (1974)). Again, the sample
file mle-advanced.inp provides an example. On the other hand, you may want to perform a check
of your analytically-computed H matrix versus a numerical approximation.
If you have a function that computes the score, this is relatively simple to do by using the fdjac
function, briefly described in section 34.3, which computes a numerical approximation to a deriva-
tive. In practice, you need a function computing g(θ) as a row vector and then use fdjac to
differentiate it numerically with respect to θ. The result can then be compared to your analytically-
computed Hessian. The code fragment below shows an example of how this can be done in the
probit case:
GMM estimation
the minimum is trivially reached at θ̂ = g −1 (X̄), since the expression in square brackets equals 0.
The above reasoning can be generalized as follows: suppose θ is an n-vector and we have m
relations like
E [fi (xt , θ)] = 0 for i = 1 . . . m, (24.3)
where E[·] is a conditional expectation on a set of p variables zt , called the instruments. In the
above simple example, m = 1 and f (xt , θ) = xt − g(θ), and the only instrument used is zt = 1.
Then, it must also be true that
h i h i
E fi (xt , θ) · zj,t = E fi,j,t (θ) = 0 for i = 1 . . . m and j = 1 . . . p; (24.4)
equation (24.4) is known as an orthogonality condition, or moment condition. The GMM estimator is
defined as the minimum of the quadratic form
223
Chapter 24. GMM estimation 224
where f̄ is a (1 × m · p) vector holding the average of the orthogonality conditions and W is some
symmetric, positive definite matrix, known as the weights matrix. A necessary condition for the
minimum to exist is the order condition n ≤ m · p.
The statistic
θ̂ = Argmin F (θ, W ) (24.6)
θ
1. Choose a positive definite W and compute the one-step GMM estimator θ̂1 . Customary choices
for W are Im·p or Im ⊗ (Z 0 Z)−1 .
2. Use θ̂1 to estimate V (fi,j,t (θ)) and use its inverse as the weights matrix. The resulting esti-
mator θ̂2 is called the two-step estimator.
3. Re-estimate V (fi,j,t (θ)) by means of θ̂2 and obtain θ̂3 ; iterate until convergence. Asymp-
totically, these extra steps are unnecessary, since the two-step estimator is consistent and
efficient; however, the iterated estimator often has better small-sample properties and should
be independent of the choice of W made at step 1.
In the special case when the number of parameters n is equal to the total number of orthogonality
conditions m · p, the GMM estimator θ̂ is the same for any choice of the weights matrix W , so the
first step is sufficient; in this case, the objective function is 0 at the minimum.
If, on the contrary, n < m · p, the second step (or successive iterations) is needed to achieve
efficiency, and the estimator so obtained can be very different, in finite samples, from the one-
step estimator. Moreover, the value of the objective function at the minimum, suitably scaled by
the number of observations, yields Hansen’s J statistic; this statistic can be interpreted as a test
statistic that has a χ 2 distribution with m · p − n degrees of freedom under the null hypothesis of
correct specification. See Davidson and MacKinnon (1993, section 17.6) for details.
In the following sections we will show how these ideas are implemented in gretl through some
examples.
E(xi ) = p · θ V (xi ) = p · θ 2
1 In section 23.2 we used a slightly different, perhaps more common, parametrization, employing θ = 1/α. We are
These will become our moment conditions; substituting the finite sample analogues of the theoret-
ical moments we have
X̄ = p̂ · θ̂ (24.7)
2
V (xi ) = p̂ · θ̂ (24.8)
V̂
Of course, the two equations above are easy to solve analytically, giving θ̂ = X̄ and p̂ = X̄ , (V̂ being
θ̂
the sample variance of xi ), but it’s instructive to see how the gmm command will solve this system
of equations numerically.
We feed gretl the necessary ingredients for GMM estimation in a command block, starting with gmm
and ending with end gmm. Three elements are compulsory within a gmm block:
orthog x ; Z
where x may be a series, matrix or list of series and Z may also be a series, matrix or list. Note the
structure of the orthogonality condition: it is assumed that the term to the left of the semicolon
represents a quantity that depends on the estimated parameters (and so must be updated in the
process of iterative estimation), while the term on the right is a constant function of the data.
The weights statement is used to specify the initial weighting matrix and its syntax is straightfor-
ward.
The params statement specifies the parameters with respect to which the GMM criterion should be
minimized; it follows the same logic and rules as in the mle and nls commands.
The minimum is found through numerical minimization via BFGS (see chapters 34 and 23). The
progress of the optimization procedure can be observed by appending the --verbose switch to
the end gmm line.
Equations 24.7 and 24.8 are not yet in the “moment condition” form required by the gmm command.
We need to transform them and arrive at something looking like E(ej,i zj,i ) = 0, with j = 1 . . . 2.
Therefore, we need two corresponding observable variables e1 and e2 with corresponding instru-
ments z1 and z2 and tell gretl that Ê(ej zj ) = 0 must be satisfied (where we used the Ê(·) notation
to indicate sample moments).
If we define the instrument as a series of ones, and set e1,i = xi − pθ, then we can re-write the first
moment condition as
Ê[(xi − pθ) · 1] = 0.
This is in the form required by the gmm command: in the required input statement “orthog e ; z”,
e will be the variable on the left (defined as a series) and z will the variable to the right of the
semicolon. Since z1,i = 1 for all i, we can use the built-in series const for that.
For the second moment condition we have, analogously,
nh i o
Ê (xi − X̄)2 − pθ 2 · 1 = 0,
Chapter 24. GMM estimation 226
so that by setting e2,i = (xi − X̄)2 − pθ 2 and z2 = z1 we can re-write the second moment condition
as Ê[e2,i · 1] = 0.
The weighting matrix, which is required by the gmm command, can be set to any 2 × 2 positive
definite matrix, since under exact identification the choice does not matter and its dimension is
determined by the number of orthogonality conditions. Therefore, we’ll use the identity matrix.
Example code follows:
#generate a gamma random variable with, say, shape p = 3 and scale theta = 2
series x = randgen(G, 3, 2)
#declare and set some initial value for parameters p and theta
scalar p = 1
scalar theta =1
gmm
scalar m = mean(x)
series e1 = x - p*theta
series e2 = (x - m)^2 - p*theta^2
orthog e1 ; const
orthog e2 ; const
weights W
params p theta
end gmm
If we want to use the unbiased estimator for the sample variance, we’d have to adjust the second
moment condition by substituting
with
Chapter 24. GMM estimation 227
One can observe tiny improvements in the point estimates, since both moved a tad closer to the
true values. This, however, is just a small-sample effect and not something you should expect in
larger samples.
• θ is β;
• the instrument is xt ;
• fi,j,t (θ) is (yt −xt β)xt = ut xt ; the orthogonality condition is interpretable as the requirement
that the regressors should be uncorrelated with the disturbances;
• W can be any symmetric positive definite matrix, since the number of parameters equals the
number of orthogonality conditions. Let’s say we choose I.
and it is easy to see why OLS and GMM coincide here: the GMM objective function has the
same minimizer as the objective function of OLS, the residual sum of squares. Note, however,
that the two functions are not equal to one another: at the minimum, F (θ, W ) = 0 while the
minimized sum of squared residuals is zero only in the special case of a perfect linear fit.
The code snippet contained in Listing 24.1 uses gretl’s gmm command to make the above opera-
tional. In example 24.1, the series e holds the “residuals” and the series x holds the regressor. If x
had been a list (a matrix), the orthog statement would have generated one orthogonality condition
for each element (column) of x.
Chapter 24. GMM estimation 228
/* initialize stuff */
series e = 0
scalar beta = 0
matrix W = I(1)
+ In cases other than one-step estimation the specified weights matrix will be overwritten with the final
weights on completion of the gmm command. If you wish to execute more than one GMM block with a
common starting-point it is therefore necessary to reinitialize the weights matrix between runs.
Partial output from this script is shown in 24.3. The estimated standard errors from GMM are
robust by default; if we supply the --robust option to the tsls command we get identical results.2
After the end gmm statement two mutually exclusive options can be specified: --two-step or
--iterate, whose meaning should be obvious.
2 The data file used in this example is available in the Stock and Watson package for gretl. See http://gretl.
sourceforge.net/gretl_data.html.
Chapter 24. GMM estimation 229
open cig_ch10.gdt
# real avg price including sales tax
genr ravgprs = avgprs / cpi
# real avg cig-specific tax
genr rtax = tax / cpi
# real average total tax
genr rtaxs = taxs / cpi
# real average sales tax
genr rtaxso = rtaxs - rtax
# logs of consumption, price, income
genr lpackpc = log(packpc)
genr lravgprs = log(ravgprs)
genr perinc = income / (pop*cpi)
genr lperinc = log(perinc)
# restrict sample to 1995 observations
smpl --restrict year==1995
# Equation (10.16) by tsls
list xlist = const lravgprs lperinc
list zlist = const rtaxso rtax lperinc
tsls lpackpc xlist ; zlist --robust
This estimator is robust with respect to heteroskedasticity, but not with respect to autocorrela-
tion. A heteroskedasticity- and autocorrelation-consistent (HAC) variant can be obtained using
the Bartlett kernel or similar. A univariate version of this is used in the context of the lrvar()
function—see equation (9.1). The multivariate version is set out in equation (24.12).
T −k k
1 XX 0
Ω̂k (θ) = wi ft (θ)ft−i (θ) , (24.12)
T t=k i=−k
Gretl computes the HAC covariance matrix by default when a GMM model is estimated on time
series data. You can control the kernel and the bandwidth (that is, the value of k in 24.12) using
the set command. See chapter 19 for further discussion of HAC estimation. You can also ask gretl
not to use the HAC version by saying
set force_hc on
Chapter 24. GMM estimation 231
where p is the asset price, U(·) is the individual’s utility function, δ is the individual’s subjective
discount rate and rt+k is the asset’s rate of return between time t and time t + k. Ft is the infor-
mation set at time t; equation (24.13) says that the utility “lost” at time t by purchasing the asset
instead of consumption goods must be matched by a corresponding increase in the (discounted)
future utility of the consumption financed by the asset’s return. Since the future is uncertain, the
individual considers his expectation, conditional on what is known at the time when the choice is
made.
We have said nothing about the nature of the asset, so equation (24.13) should hold whatever asset
we consider; hence, it is possible to build a system of equations like (24.13) for each asset whose
price we observe.
If we are willing to believe that
• the economy as a whole can be represented as a single gigantic and immortal representative
individual, and
x α −1
• the function U(x) = α is a faithful representation of the individual’s preferences,
then, setting k = 1, equation (24.13) implies the following for any asset j:
where Ct is aggregate consumption and α and δ are the risk aversion and discount rate of the
representative individual. In this case, it is easy to see that the “deep” parameters α and δ can be
estimated via GMM by using
rj,t+1 Ct+1 α−1
et = δ −1
pj,t Ct
as the moment condition, while any variable known at time t may serve as an instrument.
In the example code given in 24.4, we replicate selected portions of table 3.7 in Hall (2005). The
variable consrat is defined as the ratio of monthly consecutive real per capita consumption (ser-
vices and nondurables) for the US, and ewr is the return–price ratio of a fictitious asset constructed
by averaging all the stocks in the NYSE. The instrument set contains the constant and two lags of
each variable.
The command set force_hc on on the second line of the script has the sole purpose of replicating
the given example: as mentioned above, it forces gretl to compute the long-run variance of the
orthogonality conditions according to equation (24.11) rather than (24.12).
We run gmm four times: one-step estimation for each of two initial weights matrices, then iterative
estimation starting from each set of initial weights. Since the number of orthogonality conditions
(5) is greater than the number of estimated parameters (2), the choice of initial weights should
make a difference, and indeed we see fairly substantial differences between the one-step estimates
Chapter 24. GMM estimation 232
open hall.gdt
set force_hc on
matrix V0 = 100000*I(nelem(inst))
matrix Z = { inst }
matrix V1 = $nobs*inv(Z’Z)
gmm e = delta*ewr*consrat^(alpha-1) - 1
orthog e ; inst
weights V0
params alpha delta
end gmm
gmm e = delta*ewr*consrat^(alpha-1) - 1
orthog e ; inst
weights V1
params alpha delta
end gmm
gmm e = delta*ewr*consrat^(alpha-1) - 1
orthog e ; inst
weights V0
params alpha delta
end gmm --iterate
gmm e = delta*ewr*consrat^(alpha-1) - 1
orthog e ; inst
weights V1
params alpha delta
end gmm --iterate
Chapter 24. GMM estimation 233
Listing 24.5: Estimation of the Consumption Based Asset Pricing Model — output
(Models 1 and 2). On the other hand, iteration reduces these differences almost to the vanishing
point (Models 3 and 4).
Part of the output is given in 24.5. It should be noted that the J test leads to a rejection of the
hypothesis of correct specification. This is perhaps not surprising given the heroic assumptions
required to move from the microeconomic principle in equation (24.13) to the aggregate system
that is actually estimated.
24.7 Caveats
A few words of warning are in order: despite its ingenuity, GMM is possibly the most fragile esti-
mation method in econometrics. The number of non-obvious choices one has to make when using
GMM is high, and in finite samples each of these can have dramatic consequences on the eventual
output. Some of the factors that may affect the results are:
1. Orthogonality conditions can be written in more than one way: for example, if E(xt − µ) = 0,
then E(xt /µ − 1) = 0 holds too. It is possible that a different specification of the moment
conditions leads to different results.
2. As with all other numerical optimization algorithms, weird things may happen when the ob-
jective function is nearly flat in some directions or has multiple minima. BFGS is usually quite
good, but there is no guarantee that it always delivers a sensible solution, if one at all.
3. The 1-step and, to a lesser extent, the 2-step estimators may be sensitive to apparently trivial
details, like the re-scaling of the instruments. Different choices for the initial weights matrix
can also have noticeable consequences.
4. With time-series data, there is no hard rule on the appropriate number of lags to use when
computing the long-run covariance matrix (see section 24.5). Our advice is to go by trial and
error, since results may be greatly influenced by a poor choice. Future versions of gretl will
include more options on covariance matrix estimation.
One of the consequences of this state of things is that replicating various well-known published
studies may be extremely difficult. Any non-trivial result is virtually impossible to reproduce unless
all details of the estimation procedure are carefully recorded.
Chapter 25
25.1 Introduction
In some contexts the econometrician chooses between alternative models based on a formal hy-
pothesis test. For example, one might choose a more general model over a more restricted one if
the restriction in question can be formulated as a testable null hypothesis, and the null is rejected
on an appropriate test.
In other contexts one sometimes seeks a criterion for model selection that somehow measures the
balance between goodness of fit or likelihood, on the one hand, and parsimony on the other. The
balancing is necessary because the addition of extra variables to a model cannot reduce the degree
of fit or likelihood, and is very likely to increase it somewhat even if the additional variables are
not truly relevant to the data-generating process.
The best known such criterion, for linear models estimated via least squares, is the adjusted R 2 ,
SSR/(n − k)
R̄ 2 = 1 −
TSS/(n − 1)
where n is the number of observations in the sample, k denotes the number of parameters esti-
mated, and SSR and TSS denote the sum of squared residuals and the total sum of squares for
the dependent variable, respectively. Compared to the ordinary coefficient of determination or
unadjusted R 2 ,
SSR
R2 = 1 −
TSS
the “adjusted” calculation penalizes the inclusion of additional parameters, other things equal.
where `(θ̂) represents the maximum loglikelihood as a function of the vector of parameter esti-
mates, θ̂, and k (as above) denotes the number of “independently adjusted parameters within the
model.” In this formulation, with AIC negatively related to the likelihood and positively related to
the number of parameters, the researcher seeks the minimum AIC.
The AIC can be confusing, in that several variants of the calculation are “in circulation.” For exam-
ple, Davidson and MacKinnon (2004) present a simplified version,
AIC = `(θ̂) − k
which is just −2 times the original: in this case, obviously, one wants to maximize AIC.
In the case of models estimated by least squares, the loglikelihood can be written as
n n
`(θ̂) = − (1 + log 2π − log n) − log SSR (25.2)
2 2
235
Chapter 25. Model selection criteria 236
Some authors simplify the formula for the case of models estimated via least squares. For instance,
William Greene writes
SSR 2k
AIC = log + (25.4)
n n
This variant can be derived from (25.3) by dividing through by n and subtracting the constant
1 + log 2π . That is, writing AICG for the version given by Greene, we have
1
AICG = AIC − (1 + log 2π )
n
SSR 2k/n
AICR = e
n
Although the Akaike criterion is designed to favor parsimony, arguably it does not go far enough
in that direction. For instance, if we have two nested models with k − 1 and k parameters respec-
tively, and if the null hypothesis that parameter k equals 0 is true, in large samples the AIC will
nonetheless tend to select the less parsimonious model about 16 percent of the time (see Davidson
and MacKinnon, 2004, chapter 15).
An alternative to the AIC which avoids this problem is the Schwarz (1978) “Bayesian information
criterion” (BIC). The BIC can be written (in line with Akaike’s formulation of the AIC) as
The multiplication of k by log n in the BIC means that the penalty for adding extra parameters
grows with the sample size. This ensures that, asymptotically, one will not select a larger model
over a correctly specified parsimonious model.
A further alternative to AIC, which again tends to select more parsimonious models than AIC,
is the Hannan–Quinn criterion or HQC (Hannan and Quinn, 1979). Written consistently with the
formulations above, this is
HQC = −2`(θ̂) + 2k log log n
The Hannan–Quinn calculation is based on the law of the iterated logarithm (note that the last term
is the log of the log of the sample size). The authors argue that their procedure provides a “strongly
consistent estimation procedure for the order of an autoregression”, and that “compared to other
strongly consistent procedures this procedure will underestimate the order to a lesser degree.”
Gretl reports the AIC, BIC and HQC (calculated as explained above) for most sorts of models. The
key point in interpreting these values is to know whether they are calculated such that smaller
values are better, or such that larger values are better. In gretl, smaller values are better: one wants
to minimize the chosen criterion.
Chapter 26
26.1 Introduction
This chapter gives a brief account of the issue of correction for degrees of freedom in the context
of econometric modeling, leading up to a discussion of the policies adopted in gretl in this regard.
We also explain how to supplement the results produced automatically by gretl if you want to apply
such a correction where gretl does not, or vice versa.
The first few sections are quite basic; experts are invited to skip to section 26.5.
The intuition behind the bias in (26.1) is straightforward. First, the quantity we seek to estimate is
defined as h i
σ 2 = E (xi − µ)2
where µ = E(xi ); it is clear that if µ were observable, a perfectly good estimator would be
n
1 X
σ̄ 2 = (xi − µ)2 .
n i=1
But this is not a practical option since µ is generally unobservable and must be estimated; therefore,
we use x̄ as an estimator of µ: it is easily shown that x̄ is the least-squares estimator, and also
(assuming normality) the ML estimator. It is unbiased, but is of course subject to sampling error;
in any given sample it is highly unlikely that x̄ = µ. Given that x̄ is the least-squares estimator,
the sum of squared deviations of the xi from any value other than x̄ must be greater than the
summation in (26.1). But since µ is almost certainly not equal to x̄, the sum of squared deviations
of the xi from µ will surely be greater than the sum of squared deviations in (26.1). It follows that
the expected value of σ̂ 2 falls short of the population variance.
The proof that s 2 is indeed the unbiased estimator can be found in any good statistics textbook,
where we also learn that the magnitude n − 1 in (26.2) can be brought under a general description
as the “degrees of freedom” of the calculation at hand. (Given x̄, the n sample values provide only
n − 1 items of information since the nth value can always be deduced via the formula for x̄.)
237
Chapter 26. Degrees of freedom correction 238
where σ 2 is the variance of the error term and X is an n × k matrix of regressors. But how should
the unknown σ 2 be estimated? The ML estimator is
n
1 X 2
σ̂ 2 = û (26.3)
n i=1 i
where the û2i are squared residuals, ui = yi − Xi β. But this estimator is biased and we typically use
the unbiased counterpart
n
1 X 2
s2 = û (26.4)
n − k i=1 i
in which n − k is the number of degrees of freedom given n residuals from a regression where k
parameters are estimated.
The standard estimator of the variance of β̂ in the context of OLS is then V = s 2 (X 0 X)−1 . And the
standard errors of the individual parameter estimates, sβ̂i , being the square roots of the diagonal
elements of V , inherit a degrees of freedom correction from the estimator s 2 .
Going one step further, consider hypothesis testing in the context of OLS. Since the variance of
β̂ is unknown and must itself be estimated, the sampling distribution of the OLS coefficients is
not, strictly speaking, normal. But if the disturbances are normally distributed (besides being IID)
then, even in small samples, the parameter estimates will follow a distribution that can be specified
exactly, namely the Student t distribution with degrees of freedom equal to the value given above,
ν = n − k.
That is, besides using a df correction in computing the standard errors of the OLS coefficients,
one uses the same ν in selecting the particular distribution to which the “t-ratio”, (β̂i − β0 )/sβ̂i ,
should be referred in order to determine the marginal significance level or p-value for the null
hypothesis that βi = β0 . This is the payoff to df correction: we get test statistics that follow a known
distribution in small samples. (In big enough samples the point is moot, since the quantitative
distinction between σ̂ 2 and s 2 vanishes.)
So far, so good. Everyone expects df correction in plain OLS standard errors just as we expect
division by n − 1 in the sample variance. And users of econometric software expect that the p-
values reported for OLS coefficients will be based on the t(ν) distribution — although they are
not always sufficiently aware that the validity of such statistics in small samples depends on the
assumption of normally distributed errors.
theorems entitle us to expect that test statistics will converge to the normal distribution, or the χ 2
distribution for multivariate tests, given big enough samples.
To “correct” or not?
The question arises, should we or should we not apply a df “correction” in reporting variance
estimates and standard errors for models that depart from the classical linear specification?
The argument against applying df adjustment is that it lacks a theoretical basis: it does not pro-
duce test statistics that follow any known distribution in small samples. In addition, if parameter
estimates are obtained via ML, it makes sense to report ML estimates of variances even if these
are biased, since it is the ML quantities that are used in computing the criterion function and in
forming likelihood-ratio tests.
On the other hand, pragmatic arguments for doing df adjustment are (a) that it makes for closer
comparability between regular OLS estimates and nonlinear ones, and (b) that it provides a “pinch
of salt” in relation to small-sample results — that is, it inflates standard errors, confidence intervals
and p-values somewhat—even if it lacks rigorous justification.
Note that even for fairly small samples, the difference between the biased and unbiased estimators
30 2
in equations (26.1) and (26.2) above will be small. For example, if n = 30 then s 2 = 29 σ̂ . In
econometric modelling proper, however, the difference can be quite substantial. If n = 50 and
k = 10, the s 2 defined in (26.4) will be 50/40 = 1.25 as large as the σ̂ 2 in (26.3), and standard errors
will be about 12 percent larger.1 One can make a case for inflating the standard errors obtained
via nonlinear estimators as a precaution against taking results to be “more precise than they really
are”.
In rejoinder to the last point, one might equally say that savvy econometricians should know to
apply a discount factor (albeit an imprecise one) to small-sample estimates outside of the classical,
normal linear model—or even that they should distrust such results and insist on large samples
before making inferences. This line of thinking suggests that test statistics such as z = β̂i /σ̂β̂i
should be referred to the distribution to which they conform asymptotically — in this case N(0, 1)
for H0 : βi = 0—if and only if the conditions for appealing to asymptotic results can be considered
as met. From this point of view df adjustment may be seen as providing a false sense of security.
1 A fairly typical situation in time-series macroeconometrics would be have between 100 and 200 quarterly observa-
tions, and to be estimating up to maybe 30 parameters including lags. In this case df correction would make a difference
to standard errors on the order of 10 percent.
Chapter 26. Degrees of freedom correction 240
Another “grey area” is the class of Feasible Generalized Least Squares (FGLS) estimators — for exam-
ple, weighted least squares following the estimation of a skedastic function, or estimators designed
to handle first-order autocorrelation, such as Cochrane–Orcutt. These depart from the classical lin-
ear model, and the theoreretical basis for inference in such models is asymptotic, yet according to
econometric tradition standard errors are generally df adjusted.
Yet another awkward case: “robust” (heteroskedasticity- and/or autocorrelation-consistent) stan-
dard errors in the context of OLS. Such estimators are justified by asymptotic arguments and in
general we cannot determine their small-sample distributions. That would argue for referring the
associated test statistics to the normal distribution. But comparability with classical standard
errors pulls in the other direction. Suppose in a particular case a robust estimator produces a
standard error that is numerically indistinguishable from the classical one: if the former is referred
to the normal distribution and the latter to the t distribution, switching to robust standard errors
will give a smaller p-value for the coefficient in question, making it appear “more significant,” and
arguably this is misleading.
• The ols and tsls commands support a --no-df-corr option to suppress degrees of freedom
adjustment. In the case of Two-Stage Least Squares it’s certainly arguable that df correction
should not be performed by default, however gretl does this, largely for comparability with
other software (for example Stata’s ivreg command). But you can override the default if you
wish.
• The estimate command, for systems of equations, also supports the --no-df-corr option
when the specified estimation method is OLS or TSLS. (For other estimators supported by
gretl’s system command no df adjustment is applied by default.)
• By default gretl uses the t distribution for statistics based on robust standard errors under
OLS. However, users can specify that p-values be calculated using the standard normal distri-
bution whenever the --robust option is passed to an estimation command, by means of the
following “set” command
set robust_z on
This will print the original coefficient estimates along with asymptotic standard errors and the
associated z-scores and (two-sided) normal p-values. The converse case is left as an exercise for
the reader.
Chapter 26. Degrees of freedom correction 241
VARs
As mentioned above, Vector Autoregressions constitute a particularly awkward case, with consid-
erations of consistency of treatment pulling in two opposite directions. For that reason gretl has
adopted an “agnostic” policy in relation to such systems. We do not offer a $vcv accessor, but
instead accessors named $xtxinv (the matrix X 0 X −1 for the system as a whole) and $sigma (an
estimate of the cross-equation variance–covariance matrix, Σ). It’s then up to the user to build an
estimate of the variance matrix of the parameter estimates — call it V — should that be required.
Note that $sigma gives the Maximum Likelihood Estimator (without a degrees of freedom adjust-
ment) so if you do
(where “**” represents Kronecker product) you obtain the MLE of the variance matrix of the param-
eter estimates. But if you want the unbiased estimator you can do
to employ a suitably inflated variant of the Σ estimate. (For VARs, and also VECMs, $ncoeff gives
the number of coefficients per equation.)
The second variant above is such that the vector of standard errors produced by
matrix SE = sqrt(diag(Vu))
agrees with the standard errors printed as part of the per-equation VAR output.
A fuller example of usage of the $xtxinv accessor is given in Listing 26.1: this shows how one can
replicate the F -tests for Granger causality that are displayed by default by the var command, with
the refinement that, depending on the setting of the USE_F flag, these tests can be done using a
small sample correction as in gretl’s output or in asymptotic (χ 2 ) form.
Vector Error Correction Models are more complex than VARs in this respect, since we employ
Johansen’s variance estimator for the “β” terms. FIXME expand on this.
Chapter 26. Degrees of freedom correction 242
open denmark.gdt
list LST = LRM LRY IBO IDE
scalar p = 2 # lags in VAR
scalar USE_F = 1 # small sample correction?
k = nelem(LST)
matrix theta = vec($coeff)
matrix V = $sigma ** $xtxinv
if USE_F
scalar df = $T - $ncoeff
V *= $T/df
endif
matrix GC = zeros(k, k)
colnames(GC, LST)
rownames(GC, LST)
if USE_F
GC /= p
matrix pvals = pvalue(F, p, df, GC)
else
matrix pvals = pvalue(X, p, GC)
endif
colnames(pvals, LST)
rownames(pvals, LST)
print GC pvals
Chapter 27
In addition to the usual application of lags and differences, gretl provides fractional differencing
and various filters commonly used in macroeconomics for trend-cycle decomposition: notably the
Hodrick–Prescott filter (Hodrick and Prescott, 1997), the Baxter–King bandpass filter (Baxter and
King, 1999) and the Butterworth filter (Butterworth, 1930).
d(d + 1) 2
f (z) = 1 + dz + z + ···
2
or, more compactly,
∞
X
f (z) = 1 + ψi zi
i=1
with Qk
i=1 (d + i − 1) d+k−1
ψk = = ψk−1
k! k
The same expansion can be used with the lag operator, so that if we defined
Yt = (1 − L)0.5 Xt
genr Y = fracdiff(X,0.5)
243
Chapter 27. Time series filters 244
T TX
−1
X 2
(yt − gt )2 + λ (gt+1 − gt ) − (gt − gt−1 ) .
t=1 t=2
The first term above is the sum of squared cyclical components ct = yt − gt . The second term is a
multiple λ of the sum of squares of the trend component’s second differences. This second term
penalizes variations in the growth rate of the trend component: the larger the value of λ, the higher
is the penalty and hence the smoother the trend series.
Note that the hpfilt function in gretl produces the cyclical component, ct , of the original series.
If you want the smoothed trend you can subtract the cycle from the original:
genr ct = hpfilt(yt)
genr gt = yt - ct
Hodrick and Prescott (1997) suggest that a value of λ = 1600 is reasonable for quarterly data. The
default value in gretl is 100 times the square of the data frequency (which, of course, yields 1600
for quarterly data). The value can be adjusted using an optional second argument to hpfilt(), as
in
As of version 2018a, the hpfilt() function accepts a third, optional Boolean argument. If set to
non-zero, what is performed is the so-called one-sided version of the filter. See Section 33.10 for
further details.
To extract the component of yt that lies between the frequencies ω and ω one could apply a
bandpass filter: Zπ
ct∗ = F ∗ (ω)eiω dZ(ω)
−π
∗
where F (ω) = 1 for ω < |ω| < ω and 0 elsewhere. This would imply, in the time domain,
applying to the series a filter with an infinite number of coefficients, which is undesirable. The
Baxter and King bandpass filter applies to yt a finite polynomial in the lag operator A(L):
ct = A(L)yt
The coefficients ai are chosen such that F (ω) = A(eiω )A(e−iω ) is the best approximation to F ∗ (ω)
for a given k. Clearly, the higher k the better the approximation is, but since 2k observations have
to be discarded, a compromise is usually sought. Moreover, the filter has also other appealing
Chapter 27. Time series filters 245
theoretical properties, among which the property that A(1) = 0, so a series with a single unit root
is made stationary by application of the filter.
In practice, the filter is normally used with monthly or quarterly data to extract the “business
cycle” component, namely the component between 6 and 36 quarters. Usual choices for k are 8 or
12 (maybe higher for monthly series). The default values for the frequency bounds are 8 and 32,
and the default value for the approximation order, k, is 8. You can adjust these values using the
full form of bkfilt(), which is
bkfilt(seriesname, f1, f2, k)
where f1 and f2 represent the lower and upper frequency bounds respectively.
where
T −2 T
Σ = {2IT − (LT + L−1
T )} and M = {2IT + (LT + L−1
T )}
is a Toeplitz matrix.
The behavior of the Butterworth filter is governed by two parameters: the frequency cutoff ω? and
an integer order, n, which determines the number of coefficients used. The λ that appears in (27.1)
is tan(ω? /2)−2n . Higher values of n produce a better approximation to the ideal filter in principle
(i.e. a sharper cut between the pass-band and the stop-band) but there is a downside: with a greater
number of coefficients numerical instability may be an issue, and the influence of the initial values
in the sample may be exaggerated.
In gretl the Butterworth filter is implemented by the bwfilt() function,1 which takes three argu-
ments: the series to filter, the order n and the frequency cutoff, ω? , expressed in degrees. The
cutoff value must be greater than 0 and less than 180. This function operates as a low-pass filter;
for the high-pass variant, subtract the filtered series from the original, as in
Pollock recommends that the parameters of the Butterworth filter be tuned to the data: one should
examine the periodogram of the series in question (possibly after removal of a polynomial trend)
in search of a “dead spot” of low power between the frequencies one wishes to exclude and the
frequencies one wishes to retain. If ω? is placed in such a dead spot then the job of separation
can be done with a relatively small n, hence avoiding numerical problems. By way of illustration,
consider the periodogram for quarterly observations on new cars sales in the US,2 1975:1 to 1990:4
(the upper panel in Figure 27.1).
1 The code for this filter is based on D. S. G. Pollock’s programs IDEOLOG and DETREND. The Pascal source code for
Chapter 27. Time series filters 246
periods
64.0 10.7 5.8 4.0 3.0 2.5 2.1
300000
250000
200000
150000
100000
50000
0
0 20 40 60 80 100 120 140 160 180
degrees
3400
QNC (original data)
3200 QNC (smoothed) 1
3000
0.8
2800
0.6
2600
2400
0.4
2200 0.2
2000 0
0 π/4 π/2 3π/4 π
1800
1600
1976 1978 1980 1982 1984 1986 1988 1990
A seasonal pattern is clearly visible in the periodogram, centered at an angle of 90◦ or 4 periods.
If we set ω? = 68◦ (or thereabouts) we should be able to excise the seasonality quite cleanly using
n = 8. The result is shown in the lower panel of the Figure, along with the frequency response or
gain plot for the chosen filter. Note the smooth and reasonably steep drop-off in gain centered on
the nominal cutoff of 68◦ ≈ 3π /8.
The apparatus that supports this sort of analysis in the gretl GUI can be found under the Variable
menu in the main window: the items Periodogram and Filter. In the periodogram dialog box you
have the option of expressing the frequency axis in degrees, which is helpful when selecting a
Butterworth filter; and in the Butterworth filter dialog you have the option of plotting the frequency
response as well as the smoothed series and/or the residual or cycle.
the former is available from http://www.le.ac.uk/users/dsgp1 and the C sources for the latter were kindly made
available to us by the author.
2 This is the variable QNC from the Ramanathan data file data9-7.
Chapter 28
28.1 Introduction
Time series models are discussed in this chapter and the next two. Here we concentrate on ARIMA
models, unit root tests, and GARCH. The following chapter deals with VARs, and chapter 30 with
cointegration and error correction.
where φ(L), and θ(L) are polynomials in the lag operator, L, defined such that Ln xt = xt−n , and
t is a white noise process. The exact content of yt , of the AR polynomial φ(), and of the MA
polynomial θ(), will be explained in the following.
Mean terms
The process yt as written in equation (28.1) has, without further qualifications, mean zero. If the
model is to be applied to real data, it is necessary to include some term to handle the possibility
that yt has non-zero mean. There are two possible ways to represent processes with nonzero
mean: one is to define µt as the unconditional mean of yt , namely the central value of its marginal
distribution. Therefore, the series ỹt = yt − µt has mean 0, and the model (28.1) applies to ỹt . In
practice, assuming that µt is a linear function of some observable variables xt , the model becomes
This is sometimes known as a “regression model with ARMA errors”; its structure may be more
apparent if we represent it using two equations:
yt = xt β + u t
φ(L)ut = θ(L)t
The model just presented is also sometimes known as “ARMAX” (ARMA + eXogenous variables). It
seems to us, however, that this label is more appropriately applied to a different model: another
way to include a mean term in (28.1) is to base the representation on the conditional mean of yt ,
that is the central value of the distribution of yt given its own past. Assuming, again, that this can
be represented as a linear combination of some observable variables zt , the model would expand
to
φ(L)yt = zt γ + θ(L)t (28.3)
The formulation (28.3) has the advantage that γ can be immediately interpreted as the vector of
marginal effects of the zt variables on the conditional mean of yt . And by adding lags of zt to
247
Chapter 28. Univariate time series models 248
this specification one can estimate Transfer Function models (which generalize ARMA by adding
the effects of exogenous variable distributed across time).
Gretl provides a way to estimate both forms. Models written as in (28.2) are estimated by maximum
likelihood; models written as in (28.3) are estimated by conditional maximum likelihood. (For more
on these options see the section on “Estimation” below.)
In the special case when xt = zt = 1 (that is, the models include a constant but no exogenous
variables) the two specifications discussed above reduce to
and
φ(L)yt = α + θ(L)t (28.5)
respectively. These formulations are essentially equivalent, but if they represent one and the same
process µ and α are, fairly obviously, not numerically identical; rather
α = 1 − φ1 − . . . − φp µ
arma p q ; y
The AR and MA lag orders, p and q, can be given either as numbers or as pre-defined scalars.
The parameter µ can be dropped if necessary by appending the option --nc (“no constant”) to the
command. If estimation of (28.5) is needed, the switch --conditional must be appended to the
command, as in
arma p q ; y --conditional
Generalizing this principle to the estimation of (28.2) or (28.3), you get that
arma p q ; y const x1 x2
Ideally, the issue broached above could be made moot by writing a more general specification that
nests the alternatives; that is
φ(L) yt − xt β = zt γ + θ(L)t ; (28.6)
we would like to generalize the arma command so that the user could specify, for any estimation
method, whether certain exogenous variables should be treated as xt s or zt s, but we’re not yet at
that point (and neither are most other software packages).
Chapter 28. Univariate time series models 249
Seasonal models
A more flexible lag structure is desirable when analyzing time series that display strong seasonal
patterns. Model (28.1) can be expanded to
arma p q ; P Q ; y
where p and q represent the non-seasonal AR and MA orders, and P and Q the seasonal orders. For
example,
arma 1 1 ; 1 1 ; y
If yt is a quarterly series (and therefore s = 4), the above equation can be written more explicitly as
• One can construct a matrix containing the desired lags (positive integer values) and supply
the name of this matrix in place of p or q.
Both forms above specify an ARMA model in which AR lags 1 and 4 are used (but not 2 and 3).
This facility is available only for the non-seasonal component of the ARMA specification.
arma p d q ; y
series tmp = y
loop i=1..d
tmp = diff(tmp)
endloop
arma p q ; tmp
arma p d q ; P D Q ; y
arma 1 0 0 ; 1 1 1 ; y
where we use the sdiff function to create a seasonal difference (e.g. for quarterly data, yt − yt−4 ).
In specifying an ARIMA model with exogenous regressors we face a choice which relates back to the
discussion of the variant models (28.2) and (28.3) above. If we choose model (28.2), the “regression
model with ARMA errors”, how should this be extended to the case of ARIMA? The issue is whether
or not the differencing that is applied to the dependent variable should also be applied to the
regressors. Consider the simplest case, ARIMA with non-seasonal differencing of order 1. We may
estimate either
φ(L)(1 − L)(yt − Xt β) = θ(L)t (28.8)
or
φ(L) (1 − L)yt − Xt β = θ(L)t (28.9)
The first of these formulations can be described as a regression model with ARIMA errors, while the
second preserves the levels of the X variables. As of gretl version 1.8.6, the default model is (28.8),
in which differencing is applied to both yt and Xt . However, when using the default estimation
method (native exact ML, see below), the option --y-diff-only may be given, in which case gretl
estimates (28.9).1
Estimation
The default estimation method for ARMA models is exact maximum likelihood estimation (under
the assumption that the error term is normally distributed), using a variety of techniques: the main
algorithm for evaluating the log-likelihood is AS197 by Melard (1984). Maximization is performed
via BFGS and the score is approximated numerically. This method produces results that are di-
rectly comparable with many other software packages. The constant, and any exogenous variables,
are treated as in equation (28.2). The covariance matrix for the parameters is computed using a
numerical approximation to the Hessian at convergence.
The alternative method, invoked with the --conditional switch, is conditional maximum likeli-
hood (CML), also known as “conditional sum of squares” (see Hamilton, 1994, p. 132). This method
was exemplified in the script 12.3, and only a brief description will be given here. Given a sample of
size T , the CML method minimizes the sum of squared one-step-ahead prediction errors generated
1 Prior to gretl 1.8.6, the default model was (28.9). We changed this for the sake of consistency with other software.
Chapter 28. Univariate time series models 251
by the model for the observations t0 , . . . , T . The starting point t0 depends on the orders of the AR
polynomials in the model. The numerical maximization method used is BHHH, and the covariance
matrix is computed using a Gauss–Newton regression.
The CML method is nearly equivalent to maximum likelihood under the hypothesis of normality;
the difference is that the first (t0 − 1) observations are considered fixed and only enter the like-
lihood function as conditioning variables. As a consequence, the two methods are asymptotically
equivalent under standard conditions— except for the fact, discussed above, that our CML imple-
mentation treats the constant and exogenous variables as per equation (28.3).
The two methods can be compared as in the following example
open data10-1
arma 1 1 ; r
arma 1 1 ; r --conditional
which produces the estimates shown in Table 28.1. As you can see, the estimates of φ and θ
are quite similar. The reported constants differ widely, as expected — see the discussion following
equations (28.4) and (28.5). However, dividing the CML constant by 1 − φ we get 7.38, which is not
far from the ML estimate of 6.93.
Parameter ML CML
µ 6.93042 (0.923882) 1.07322 (0.488661)
φ 0.855360 (0.0511842) 0.852772 (0.0450252)
θ 0.588056 (0.0986096) 0.591838 (0.0456662)
1. Estimate a pure AR model by Least Squares (nonlinear least squares if the model requires
it, otherwise OLS). Set the AR parameter values based on this regression and set the MA
parameters to a small positive value (0.0001).
2. The Hannan–Rissanen method: First estimate an autoregressive model by OLS and save the
residuals. Then in a second OLS pass add appropriate lags of the first-round residuals to the
model, to obtain estimates of the MA parameters.
To see the details of the ARMA estimation procedure, add the --verbose option to the command.
This prints a notice of the initialization method used, as well as the parameter values and log-
likelihood at each iteration.
Besides the built-in initialization mechanisms, the user has the option of specifying a set of starting
values manually. This is done via the set command: the first argument should be the keyword
initvals and the second should be the name of a pre-specified matrix containing starting values.
For example
The specified matrix should have just as many parameters as the model: in the example above
there are three parameters, since the model implicitly includes a constant. The constant, if present,
is always given first; otherwise the order in which the parameters are expected is the same as the
order of specification in the arma or arima command. In the example the constant is set to zero,
φ1 to 0.85, and θ1 to 0.34.
You can get gretl to revert to automatic initialization via the command set initvals auto.
Two variants of the BFGS algorithm are available in gretl. In general we recommend the default vari-
ant, which is based on an implementation by Nash (1990), but for some problems the alternative,
limited-memory version (L-BFGS-B, see Byrd et al., 1995) may increase the chances of convergence
on the ML solution. This can be selected via the --lbfgs option to the arma command.
arma p q ; y --x-12-arima
As with native estimation, the default is to use exact ML but there is the option of using conditional
ML with the --conditional flag. However, please note that when X-12-ARIMA is used in conditional
ML mode, the comments above regarding the variant treatments of the mean of the process yt do
not apply. That is, when you use X-12-ARIMA the model that is estimated is (28.2), regardless
of whether estimation is by exact ML or conditional ML. In addition, the treatment of exogenous
regressors in the context of ARIMA differencing is always that shown in equation (28.8).
Forecasting
ARMA models are often used for forecasting purposes. The autoregressive component, in particu-
lar, offers the possibility of forecasting a process “out of sample” over a substantial time horizon.
Gretl supports forecasting on the basis of ARMA models using the method set out by Box and
Jenkins (1976).2 The Box and Jenkins algorithm produces a set of integrated AR coefficients which
take into account any differencing of the dependent variable (seasonal and/or non-seasonal) in the
ARIMA context, thus making it possible to generate a forecast for the level of the original variable.
By contrast, if you first difference a series manually and then apply ARMA to the differenced series,
forecasts will be for the differenced series, not the level. This point is illustrated in Listing 28.1.
The parameter estimates are identical for the two models. The forecasts differ but are mutually
consistent: the variable fcdiff emulates the ARMA forecast (static, one step ahead within the
sample range, and dynamic out of sample).
open greene18_2.gdt
# log of quarterly U.S. nominal GNP, 1950:1 to 1983:4
series y = log(Y)
# and its first difference
series dy = diff(y)
# reserve 2 years for out-of-sample forecast
smpl ; 1981:4
# Estimate using ARIMA
arima 1 1 1 ; y
# forecast over full period
smpl --full
fcast fc1
# Return to sub-sample and run ARMA on the first difference of y
smpl ; 1981:4
arma 1 1 ; dy
smpl --full
fcast fc2
series fcdiff = (t<=1982:1)? (fc1 - y(-1)) : (fc1 - fc1(-1))
# compare the forecasts over the later period
smpl 1981:1 1983:4
print y fc1 fc2 fcdiff --byobs
This test statistic is probably the best-known and most widely used unit root test. It is a one-sided
test whose null hypothesis is ϕ = 0 versus the alternative ϕ < 0 (and hence large negative values
of the test statistic lead to the rejection of the null). Under the null, yt must be differenced at least
once to achieve stationarity; under the alternative, yt is already stationary and no differencing is
required.
One peculiar aspect of this test is that its limit distribution is non-standard under the null hy-
pothesis: moreover, the shape of the distribution, and consequently the critical values for the test,
depends on the form of the µt term. A full analysis of the various cases is inappropriate here:
Hamilton (1994) contains an excellent discussion, but any recent time series textbook covers this
topic. Suffice it to say that gretl allows the user to choose the specification for µt among four
different alternatives:
µt command option
0 --nc
µ0 --c
µ0 + µ1 t --ct
2
µ0 + µ1 t + µ1 t --ctt
These option flags are not mutually exclusive; when they are used together the statistic will be
reported separately for each selected case. By default, gretl uses the combination --c --ct. For
each case, approximate p-values are calculated by means of the algorithm developed in MacKinnon
(1996).
The gretl command used to perform the test is adf; for example
adf 4 x1
would compute the test statistic as the t-statistic for ϕ in equation 28.10 with p = 4 in the two
cases µt = µ0 and µt = µ0 + µ1 t.
The number of lags (p in equation 28.10) should be chosen as to ensure that (28.10) is a parametriza-
tion flexible enough to represent adequately the short-run persistence of ∆yt . Setting p too low
results in size distortions in the test, whereas setting p too high leads to low power. As a conve-
nience to the user, the parameter p can be automatically determined. Setting p to a negative num-
ber triggers a sequential procedure that starts with p lags and decrements p until the t-statistic for
the parameter γp exceeds 1.645 in absolute value.
a time; by default the constant-only model is used but a trend can be added using the --ct flag.
When a trend is present in this test MacKinnon-type p-values are not available; instead we show
critical values from Table 1 in Elliott et al. (1996).
kpss m y
where m is an integer representing the bandwidth or window size used in the formula for estimating
the long run variance:
m
|i|
X
σ̄ 2 = 1− γ̂i
i=−m
m+1
The γ̂i terms denote the empirical autocovariances of et from order −m through m. For this
estimator to be consistent, m must be large enough to accommodate the short-run persistence of
et , but not too large compared to the sample size T . If the supplied m is non-positive a default value
1/4
T
is computed, namely the integer part of 4 100 .
The above concept can be generalized to the case where yt is thought to be stationary around a
deterministic trend. In this case, formula (28.11) remains unchanged, but the series et is defined as
the residuals from an OLS regression of yt on a constant and a linear trend. This second form of
the test is obtained by appending the --trend option to the kpss command:
kpss n y --trend
Note that in this case the asymptotic distribution of the test is different and the critical values
reported by gretl differ accordingly.
pi
X
∆yit = µit + ϕi yi,t−1 + γij ∆yi,t−j + it (28.12)
j=1
The model (28.12) allows for maximal heterogeneity across the individuals in the panel: the param-
eters of the deterministic term, the autoregressive coefficient ϕ, and the lag order p are all specific
to the individual, indexed by i.
One possible modification of this model is to impose the assumption that ϕi = ϕ for all i; that is,
the individual time series share a common autoregressive root (although they may differ in respect
of other statistical properties). The choice of whether or not to impose this assumption has an
important bearing on the hypotheses under test. Under model (28.12) the joint null is ϕi = 0 for
all i, meaning that all the individual time series are non-stationary, and the alternative (simply the
negation of the null) is that at least one individual time series is stationary. When a common ϕ is
assumed, the null is that ϕ = 0 and the alternative is that ϕ < 0. The null still says that all the
individual series are non-stationary, but the alternative now says that they are all stationary. The
choice of model should take this point into account, as well as the gain in power from forming a
pooled estimate of ϕ and, of course, the plausibility of assuming a common AR(1) coefficient.3
In gretl, the formulation (28.12) is used automatically when the adf command is used on panel
data. The joint test statistic is formed using the method of Im, Pesaran and Shin (2003). In this
context the behavior of adf differs from regular time-series data: only one case of the deterministic
term is handled per invocation of the command; the default is that µit includes just a constant but
the --nc and --ct flags can be used to suppress the constant or to include a trend, respectively;
and the quadratic trend option --ctt is not available.
The alternative that imposes a common value of ϕ is implemented via the levinlin command.
The test statistic is computed as per Levin, Lin and Chu (2002). As with the adf command, the first
argument is the lag order and the second is the name of the series to test; and the default case for
the deterministic component is a constant only. The options --nc and --ct have the same effect
as with adf. One refinement is that the lag order may be given in either of two forms: if a scalar
is given, this is taken to represent a common value of p for all individuals, but you may instead
provide a vector holding a set of pi values, hence allowing the order of autocorrelation of the series
to differ by individual. So, for example, given
levinlin 2 y
levinlin {2,2,3,3,4,4} y
the first command runs a joint ADF test with a common lag order of 2, while the second (which
assumes a panel with six individuals) allows for differing short-run dynamics. The first argument
to levinlin can be given as a set of comma-separated integers enclosed in braces, as shown above,
or as the name of an appropriately dimensioned pre-defined matrix (see chapter 16).
Besides variants of the ADF test, the KPSS test also can be used with panel data via the kpss
command. In this case the test (of the null hypothesis that the given time series is stationary for
all individuals) is implemented using the method of Choi (2001). This is an application of meta-
analysis, the statistical technique whereby an overall or composite p-value for the test of a given
null hypothesis can be computed from the p-values of a set of separate tests. Unfortunately, in
the case of the KPSS test we are limited by the unavailability of precise p-values, although if an
individual test statistic falls between the 10 percent and 1 percent critical values we are able to
interpolate with a fair degree of confidence. This gives rise to four cases.
1. All the individual KPSS test statistics fall between the 10 percent and 1 percent critical values:
the Choi method gives us a plausible composite p-value.
3 If the assumption of a common ϕ seems excessively restrictive, bear in mind that we routinely assume common
slope coefficients when estimating panel models, even if this is unlikely to be literally true.
Chapter 28. Univariate time series models 257
2. Some of the KPSS test statistics exceed the 1 percent value and none fall short of the 10
percent value: we can give an upper bound for the composite p-value by setting the unknown
p-values to 0.01.
3. Some of the KPSS test statistics fall short of the 10 percent critical value but none exceed the
1 percent value: we can give a lower bound to the composite p-value by setting the unknown
p-values to 0.10.
4. None of the above conditions are satisfied: the Choi method fails to produce any result for
the composite KPSS test.
2. Run a “cointegrating regression” via OLS. For this we select one of the potentially cointegrated
variables as dependent, and include the other potentially cointegrated variables as regressors.
The idea is that cointegration is supported if (a) the null of non-stationarity is not rejected for each
of the series individually, in step 1, while (b) the null is rejected for the residuals at step 3. That is,
each of the individual series is I(1) but some linear combination of the series is I(0).
This test is implemented in gretl by the coint command, which requires an integer lag order
(for the ADF tests) followed by a list of variables to be tested, the first of which will be taken
as dependent in the cointegrating regression. Please see the online help for coint, or the Gretl
Command Reference, for further details.
where the εt s are independently and identically distributed (iid) with mean zero and variance 1,
and where σt is taken to be the positive square root of σt2 . Ωt−1 denotes the information set as of
time t − 1 and σt2 is the conditional variance: that is, the variance conditional on information dated
t − 1 and earlier.
It is important to notice the difference between ARCH and an ordinary autoregressive error process.
The simplest (first-order) case of the latter can be written as
where the εt s are independently and identically distributed with mean zero and variance σ 2 . With
an AR(1) error, if ρ is positive then a positive value of ut will tend to be followed by a positive
ut+1 . With an ARCH error process, a disturbance ut of large absolute value will tend to be followed
by further large absolute values, but with no presumption that the successive values will be of the
same sign. ARCH in asset prices is a “stylized fact” and is consistent with market efficiency; on the
other hand autoregressive behavior of asset prices would violate market efficiency.
One can test for ARCH of order q in the following way:
1. Estimate the model of interest via OLS and save the squared residuals, û2t .
2. Perform an auxiliary regression in which the current squared residual is regressed on a con-
stant and q lags of itself.
3. Find the T R 2 value (sample size times unadjusted R 2 ) for the auxiliary regression.
4. Refer the T R 2 value to the χ 2 distribution with q degrees of freedom, and if the p-value is
“small enough” reject the null hypothesis of homoskedasticity in favor of the alternative of
ARCH(q).
This test is implemented in gretl via the modtest command with the --arch option, which must
follow estimation of a time-series model by OLS (either a single-equation model or a VAR). For
example,
ols y 0 x
modtest 4 --arch
This example specifies an ARCH order of q = 4; if the order argument is omitted, q is set equal to
the periodicity of the data. In the graphical interface, the ARCH test is accessible from the “Tests”
menu in the model window (again, for single-equation OLS or VARs).
GARCH
The simple ARCH(q) process is useful for introducing the general concept of conditional het-
eroskedasticity in time series, but it has been found to be insufficient in empirical work. The
dynamics of the error variance permitted by ARCH(q) are not rich enough to represent the patterns
found in financial data. The generalized ARCH or GARCH model is now more widely used.
The representation of the variance of a process in the GARCH model is somewhat (but not exactly)
analogous to the ARMA representation of the level of a time series. The variance at time t is allowed
to depend on both past values of the variance and past values of the realized squared disturbance,
as shown in the following system of equations:
yt = Xt β + ut (28.13)
ut = σt εt (28.14)
q p
X X
σt2 = α0 + αi u2t−i + 2
δj σt−j (28.15)
i=1 j=1
As above, εt is an iid sequence with unit variance. Xt is a matrix of regressors (or in the simplest
case, just a vector of 1s allowing for a non-zero mean of yt ). Note that if p = 0, GARCH collapses
to ARCH(q): the generalization is embodied in the δj terms that multiply previous values of the
error variance.
In principle the underlying innovation, εt , could follow any suitable probability distribution, and
besides the obvious candidate of the normal or Gaussian distribution the Student’s t distribution
has been used in this context. Currently gretl only handles the case where εt is assumed to be
Gaussian. However, when the --robust option to the garch command is given, the estimator gretl
Chapter 28. Univariate time series models 259
uses for the covariance matrix can be considered Quasi-Maximum Likelihood even with non-normal
disturbances. See below for more on the options regarding the GARCH covariance matrix.
Example:
garch p q ; y const x
where p ≥ 0 and q > 0 denote the respective lag orders as shown in equation (28.15). These values
can be supplied in numerical form or as the names of pre-defined scalar variables.
GARCH estimation
Estimation of the parameters of a GARCH model is by no means a straightforward task. (Consider
equation 28.15: the conditional variance at any point in time, σt2 , depends on the conditional
variance in earlier periods, but σt2 is not observed, and must be inferred by some sort of Maximum
Likelihood procedure.) By default gretl uses native code that employs the BFGS maximizer; you
also have the option (activated by the --fcp command-line switch) of using the method proposed
by Fiorentini et al. (1996),4 which was adopted as a benchmark in the study of GARCH results
by McCullough and Renfro (1998). It employs analytical first and second derivatives of the log-
likelihood, and uses a mixed-gradient algorithm, exploiting the information matrix in the early
iterations and then switching to the Hessian in the neighborhood of the maximum likelihood. (This
progress can be observed if you append the --verbose option to gretl’s garch command.)
Several options are available for computing the covariance matrix of the parameter estimates in
connection with the garch command. At a first level, one can choose between a “standard” and a
“robust” estimator. By default, the Hessian is used unless the --robust option is given, in which
case the QML estimator is used. A finer choice is available via the set command, as shown in
Table 28.2.
command effect
set garch_vcv hessian Use the Hessian
set garch_vcv im Use the Information Matrix
set garch_vcv op Use the Outer Product of the Gradient
set garch_vcv qml QML estimator
set garch_vcv bw Bollerslev–Wooldridge “sandwich” estimator
It is not uncommon, when one estimates a GARCH model for an arbitrary time series, to find that
the iterative calculation of the estimates fails to converge. For the GARCH model to make sense,
there are strong restrictions on the admissible parameter values, and it is not always the case
that there exists a set of values inside the admissible parameter space for which the likelihood is
maximized.
The restrictions in question can be explained by reference to the simplest (and much the most
common) instance of the GARCH model, where p = q = 1. In the GARCH(1, 1) model the conditional
variance is
σt2 = α0 + α1 u2t−1 + δ1 σt−1
2
(28.16)
Taking the unconditional expectation of (28.16) we get
σ 2 = α0 + α1 σ 2 + δ1 σ 2
4 The algorithm is based on Fortran code deposited in the archive of the Journal of Applied Econometrics by the
so that
α0
σ2 =
1 − α1 − δ 1
For this unconditional variance to exist, we require that α1 + δ1 < 1, and for it to be positive we
require that α0 > 0.
A common reason for non-convergence of GARCH estimates (that is, a common reason for the non-
existence of αi and δi values that satisfy the above requirements and at the same time maximize
the likelihood of the data) is misspecification of the model. It is important to realize that GARCH, in
itself, allows only for time-varying volatility in the data. If the mean of the series in question is not
constant, or if the error process is not only heteroskedastic but also autoregressive, it is necessary
to take this into account when formulating an appropriate model. For example, it may be necessary
to take the first difference of the variable in question and/or to add suitable regressors, Xt , as in
(28.13).
Chapter 29
Vector Autoregressions
Gretl provides a standard set of procedures for dealing with the multivariate time-series models
known as VARs (Vector AutoRegression). More general models — such as VARMAs, nonlinear models
or multivariate GARCH models—are not provided as of now, although it is entirely possible to
estimate them by writing custom procedures in the gretl scripting language. In this chapter, we will
briefly review gretl’s VAR toolbox.
29.1 Notation
A VAR is a structure whose aim is to model the time persistence of a vector of n time series, yt ,
via a multivariate autoregression, as in
The number of lags p is called the order of the VAR. The vector xt , if present, contains a set of
exogenous variables, often including a constant, possibly with a time trend and seasonal dummies.
The vector t is typically assumed to be a vector white noise, with covariance matrix Σ.
Equation (29.1) can be written more compactly as
∂yt
Θi = (29.4)
∂t−i
The Θi matrices may be derived by recursive substitution in equation (29.1): for example, assuming
for simplicity that B = 0 and p = 1, equation (29.1) would become
yt = Ayt−1 + t
261
Chapter 29. Vector Autoregressions 262
29.2 Estimation
The gretl command for estimating a VAR is var which, in the command line interface, is invoked
in the following manner:
where p is a scalar (the VAR order) and Ylist is a list of variables specifying the content of yt .
The optional Xlist argument can be used to specify a set of exogenous variables. If this argument
is omitted, the vector xt is taken to contain a constant (only); if present, it must be separated
from Ylist by a semicolon. Note, however, that a few common choices can be obtained in a
simpler way: the options --trend and --seasonals call for inclusion of a linear trend and a set of
seasonal dummies respectively. In addition the --nc option (no constant) can be used to suppress
the standard inclusion of a constant.
The “<-” construct can be used to store the model under a name (see section 3.2), if so desired. To
estimate a VAR using the graphical interface, choose “Time Series, Vector Autoregression”, under
the Model menu.
The parameters in eq. (29.1) are typically free from restrictions, which implies that multivariate
OLS provides a consistent and asymptotically efficient estimator of all the parameters.1 Given
the simplicity of OLS, this is what every software package, including gretl, uses; example script
29.1 exemplifies the fact that the var command gives you exactly the output you would have
from a battery of OLS regressions. The advantage of using the dedicated command is that, after
estimation is done, it makes it much easier to access certain quantities and manage certain tasks.
For example, the $coeff accessor returns the estimated coefficients as a matrix with n columns
and $sigma returns an estimate of the matrix Σ, the covariance matrix of t .
Moreover, for each variable in the system an F test is automatically performed, in which the null hy-
pothesis is that no lags of variable j are significant in the equation for variable i. This is commonly
known as a Granger causality test.
In addition, two accessors become available for the companion matrix ($compan) and the VMA rep-
resentation ($vma). The latter deserves a detailed description: since the VMA representation (29.5)
is of infinite order, gretl defines a horizon up to which the Θi matrices are computed automatically.
1 In fact, under normality of OLS is indeed the conditional ML estimator. You may want to use other methods if you
t
need to estimate a VAR in which some parameters are constrained.
Chapter 29. Vector Autoregressions 263
scalar p = 2
list X = LHUR infl
list Xlag = lags(p,X)
loop foreach i X
ols $i const Xlag
endloop
var p X
...
Equation 1: LHUR
Periodicity horizon
Quarterly 20 (5 years)
Monthly 24 (2 years)
Daily 3 weeks
All other cases 10
By default, this is a function of the periodicity of the data (see table 29.1), but it can be set by the
user to any desired value via the set command with the horizon parameter, as in
set horizon 30
Calling the horizon h, the $vma accessor returns an (h + 1) × n2 matrix, in which the (i + 1)-th row
is the vectorized form of Θi .
When the --lagselect option is given, estimation is performed for all lags up to p and a table
is printed: it displays, for each order, a Likelihood Ratio test for the order p versus p − 1, plus
an array of information criteria (see chapter 25). For each information criterion in the table, a
star indicates what appears to be the “best” choice. The same output can be obtained through the
graphical interface via the “Time Series, VAR lag selection” entry under the Model menu.
Warning: in finite samples the choice of the maximum lag, p, may affect the outcome of the proce-
dure. This is not a bug, but rather an unavoidable side effect of the way these comparisons should
be made. If your sample contains T observations and you invoke the lag selection procedure with
maximum order p, gretl examines all VARs of order ranging form 1 to p, estimated on a uniform
sample of T − p observations. In other words, the comparison procedure does not use all the avail-
able data when estimating VARs of order less than p, so as to ensure that all the models in the
comparison are estimated on the same data range. Choosing a different value of p may therefore
alter the results, although this is unlikely to happen if your sample size is reasonably large.
An example of this unpleasant phenomenon is given in example script 29.2. As can be seen, ac-
cording to the Hannan-Quinn criterion, order 2 seems preferable to order 1 if the maximum tested
order is 4, but the situation is reversed if the maximum tested order is 6.
The asterisks below indicate the best (that is, minimized) values
of the respective information criteria, AIC = Akaike criterion,
BIC = Schwarz Bayesian criterion and HQC = Hannan-Quinn criterion.
The asterisks below indicate the best (that is, minimized) values
of the respective information criteria, AIC = Akaike criterion,
BIC = Schwarz Bayesian criterion and HQC = Hannan-Quinn criterion.
∂yt
Ck = = Θk K, (29.6)
∂ut−i
known as the structural VMA representation. From the Ck matrices defined in equation (29.6) two
quantities of interest may be derived: the Impulse Response Function (IRF) and the Forecast Error
Variance Decomposition (FEVD).
The IRF of variable i to shock j is simply the sequence of the elements in row i and column j of
the Ck matrices. In symbols:
∂yi,t
Ii,j,k =
∂uj,t−k
As a rule, Impulse Response Functions are plotted as a function of k, and are interpreted as the
effect that a shock has on an observable variable through time. Of course, what we observe are
the estimated IRFs, so it is natural to endow them with confidence intervals: following common
practice, gretl computes the confidence intervals by using the bootstrap;2 details are given later in
this section.
Another quantity of interest that may be computed from the structural VMA representation is the
Forecast Error Variance Decomposition (FEVD). The forecast error variance after h steps is given by
h
X
Ωh = Ck Ck0
k=0
where k ci.l is, trivially, the i, l element of Ck . As a consequence, the share of uncertainty on variable
i that can be attributed to the j-th shock after h periods equals
Ph 2
k=0 (k ci.j )
V Di,j,h = Ph Pn .
2
k=0 l=1 (k ci.l )
This makes it possible to quantify which shocks are most important to determine a certain variable
in the short and/or in the long run.
Triangularization
The formula 29.6 takes K as known, while of course it has to be estimated. The estimation problem
has been the subject of an enormous body of literature we will not even attempt to summarize
here: see for example (Lütkepohl, 2005, chapter 9).
Suffice it to say that the most popular choice dates back to Sims (1980), and consists in assuming
that K is lower triangular, so its estimate is simply the Cholesky decomposition of the estimate of Σ.
The main consequence of this choice is that the ordering of variables within the vector yt becomes
meaningful: since K is also the matrix of Impulse Response Functions at lag 0, the triangularity
2 It is possible, in principle, to compute analytical confidence intervals via an asymptotic approximation, but this is
not a very popular choice: asymptotic formulae are known to often give a very poor approximation of the finite-sample
properties.
Chapter 29. Vector Autoregressions 267
assumption means that the first variable in the ordering responds instantaneously only to shock
number 1, the second one only to shocks 1 and 2, and so forth. For this reason, each variable is
thought to “own” one shock: variable 1 owns shock number 1, and so on.
In this sort of exercise, therefore, the ordering of the y variables is important, and the applied
literature has developed the “most exogenous first” mantra — where, in this setting, “exogenous”
really means “instantaneously insensitive to structural shocks”.3 To put it differently, if variable
foo comes before variable bar in the Y list, it follows that the shock owned by foo affects bar
instantaneously, but not vice versa.
Impulse Response Functions and the FEVD can be printed out via the command line interface by us-
ing the --impulse-response and --variance-decomp options, respectively. If you need to store
them into matrices, you can compute the structural VMA and proceed from there. For example, the
following code snippet shows you how to compute a matrix containing the IRFs:
open denmark
list Y = 1 2 3 4
scalar n = nelem(Y)
var 2 Y --quiet --impulse
matrix K = cholesky($sigma)
matrix V = $vma
matrix IRF = V * (K ** I(n))
print IRF
IRF bootstrap
FIXME: todo
impacts on every variable after one lag, so nothing is really exogenous here. A better choice of words would probably
have been something like “sturdy”, but it’s too late now.
Chapter 29. Vector Autoregressions 268
version a single test is run for each equation using just the specified lag order. The example also
exposes what exactly is returned by the $test and $pvalue accessors in the two variants.
Equation 1:
Ljung-Box Q’ = 6.11506 with p-value = P(Chi-square(4) > 6.11506) = 0.191
Equation 2:
Ljung-Box Q’ = 1.67136 with p-value = P(Chi-square(4) > 1.67136) = 0.796
Equation 3:
Ljung-Box Q’ = 1.59931 with p-value = P(Chi-square(4) > 1.59931) = 0.809
30.1 Introduction
The twin concepts of cointegration and error correction have drawn a good deal of attention in
macroeconometrics over recent years. The attraction of the Vector Error Correction Model (VECM)
is that it allows the researcher to embed a representation of economic equilibrium relationships
within a relatively rich time-series specification. This approach overcomes the old dichotomy be-
tween (a) structural models that faithfully represented macroeconomic theory but failed to fit the
data, and (b) time-series models that were accurately tailored to the data but difficult if not impos-
sible to interpret in economic terms.
The basic idea of cointegration relates closely to the concept of unit roots (see section 28.3). Sup-
pose we have a set of macroeconomic variables of interest, and we find we cannot reject the hypoth-
esis that some of these variables, considered individually, are non-stationary. Specifically, suppose
we judge that a subset of the variables are individually integrated of order 1, or I(1). That is, while
they are non-stationary in their levels, their first differences are stationary. Given the statistical
problems associated with the analysis of non-stationary data (for example, the threat of spurious
regression), the traditional approach in this case was to take first differences of all the variables
before proceeding with the analysis.
But this can result in the loss of important information. It may be that while the variables in
question are I(1) when taken individually, there exists a linear combination of the variables that
is stationary without differencing, or I(0). (There could be more than one such linear combina-
tion.) That is, while the ensemble of variables may be “free to wander” over time, nonetheless the
variables are “tied together” in certain ways. And it may be possible to interpret these ties, or
cointegrating vectors, as representing equilibrium conditions.
For example, suppose we find some or all of the following variables are I(1): money stock, M, the
price level, P , the nominal interest rate, R, and output, Y . According to standard theories of the
demand for money, we would nonetheless expect there to be an equilibrium relationship between
real balances, interest rate and output; for example
m − p = γ0 + γ1 y + γ2 r γ1 > 0, γ2 < 0
m − p − γ1 y − γ2 r = γ0
Realistically, we should not expect this condition to be satisfied each period. We need to allow for
the possibility of short-run disequilibrium. But if the system moves back towards equilibrium fol-
lowing a disturbance, it follows that the vector x = (m, p, y, r )0 is bound by a cointegrating vector
β0 = (β1 , β2 , β3 , β4 ), such that β0 x is stationary (with a mean of γ0 ). Furthermore, if equilibrium is
correctly characterized by the simple model above, we have β2 = −β1 , β3 < 0 and β4 > 0. These
things are testable within the context of cointegration analysis.
There are typically three steps in this sort of analysis:
1. Test to determine the number of cointegrating vectors, the cointegrating rank of the system.
2. Estimate a VECM with the appropriate rank, but subject to no further restrictions.
269
Chapter 30. Cointegration and Vector Error Correction Models 270
The following sections expand on each of these points, giving further econometric details and
explaining how to implement the analysis using gretl.
But since yt−i ≡ yt−1 − (∆yt−1 + ∆yt−2 + · · · + ∆yt−i+1 ), we can re-write the above as
p−1
X
∆yt = µt + Πyt−1 + Γi ∆yt−i + t , (30.2)
i=1
Pp Pp
where Π = i=1 Ai − I and Γi = − j=i+1 Aj . This is the VECM representation of (30.1).
The interpretation of (30.2) depends crucially on r , the rank of the matrix Π.
• Cointegration occurs in between, when 0 < r < n and Π can be written as αβ0 . In this case,
yt is I(1), but the combination zt = β0 yt is I(0). If, for example, r = 1 and the first element
of β was −1, then one could write zt = −y1,t + β2 y2,t + · · · + βn yn,t , which is equivalent to
saying that
y1t = β2 y2,t + · · · + βn yn,t − zt
is a long-run equilibrium relationship: the deviations zt may not be 0 but they are stationary.
In this case, (30.2) can be written as
p−1
X
∆yt = µt + αβ0 yt−1 + Γi ∆yt−i + t . (30.3)
i=1
If β were known, then zt would be observable and all the remaining parameters could be
estimated via OLS. In practice, the procedure estimates β first and then the rest.
The rank of Π is investigated by computing the eigenvalues of a closely related matrix whose rank
is the same as Π: however, this matrix is by construction symmetric and positive semidefinite. As a
consequence, all its eigenvalues are real and non-negative, and tests on the rank of Π can therefore
be carried out by testing how many eigenvalues are 0.
If all the eigenvalues are significantly different from 0, then all the processes are stationary. If,
on the contrary, there is at least one zero eigenvalue, then the yt process is integrated, although
some linear combination β0 yt might be stationary. At the other extreme, if no eigenvalues are
significantly different from 0, then not only is the process yt non-stationary, but the same holds
for any linear combination β0 yt ; in other words, no cointegration occurs.
Estimation typically proceeds in two stages: first, a sequence of tests is run to determine r , the
cointegration rank. Then, for a given rank the parameters in equation (30.3) are estimated. The two
commands that gretl offers for estimating these systems are coint2 and vecm, respectively.
The syntax for coint2 is
Chapter 30. Cointegration and Vector Error Correction Models 271
where p is the number of lags in (30.1); ylist is a list containing the yt variables; xlist is an
optional list of exogenous variables; and zlist is another optional list of exogenous variables
whose effects are assumed to be confined to the cointegrating relationships.
The syntax for vecm is
where p is the number of lags in (30.1); r is the cointegration rank; and the lists ylist, xlist and
zlist have the same interpretation as in coint2.
Both commands can be given specific options to handle the treatment of the deterministic compo-
nent µt . These are discussed in the following section.
µt = µ0 + µ1 · t.
In order to have the model mimic as closely as possible the features of the observed data, there is a
preliminary question to settle. Do the data appear to follow a deterministic trend? If so, is it linear
or quadratic?
Once this is established, one should impose restrictions on µ0 and µ1 that are consistent with this
judgement. For example, suppose that the data do not exhibit a discernible trend. This means that
∆yt is on average zero, so it is reasonable to assume that its expected value is also zero. Write
equation (30.2) as
Γ (L)∆yt = µ0 + µ1 · t + αzt−1 + t , (30.4)
where zt = β0 yt is assumed to be stationary and therefore to possess finite moments. Taking
unconditional expectations, we get
0 = µ0 + µ1 · t + αmz .
Since the left-hand side does not depend on t, the restriction µ1 = 0 is a safe bet. As for µ0 , there are
just two ways to make the above expression true: either µ0 = 0 with mz = 0, or µ0 equals −αmz .
The latter possibility is less restrictive in that the vector µ0 may be non-zero, but is constrained to
be a linear combination of the columns of α. In that case, µ0 can be written as α · c, and one may
write (30.4) as " #
0 yt−1
Γ (L)∆yt = α β c + t .
1
The long-run relationship therefore contains an intercept. This type of restriction is usually written
0
α⊥ µ0 = 0,
yt = k + xt + ut
where, again, k is a real number and ut is a white noise process. Since ut is stationary by definition,
xt and yt cointegrate: that is, their difference
zt = yt − xt = k + ut
is a stationary process. For k = 0, zt is simple zero-mean white noise, whereas for k 6= 0 the process
zt is white noise with a non-zero mean.
After some simple substitutions, the two equations above can be represented jointly as a VAR(1)
system " # " # " #" # " #
yt k+m 0 1 yt−1 ut + εt
= + +
xt m 0 1 xt−1 εt
or in VECM form
" # " # " #" # " #
∆yt k+m −1 1 yt−1 ut + εt
= + + =
∆xt m 0
0 xt−1 εt
" # " # " # " #
k+m −1 h i yt−1 ut + εt
= + 1 −1 + =
m 0 xt−1 εt
" #
0 yt−1
= µ0 + αβ + ηt = µ0 + αzt−1 + ηt ,
xt−1
1. m 6= 0: In this case xt is trended, as we just saw; it follows that yt also follows a linear trend
because on average it keeps at a fixed distance k from xt . The vector µ0 is unrestricted.
which is not null and therefore the VECM shown above does have a constant term. The
constant, however, is subject to the restriction that its second element must be 0. More
generally, µ0 is a multiple of the vector α. Note that the VECM could also be written as
" # " # yt−1 " #
∆yt −1 h i ut + εt
= 1 −1 −k xt−1 +
∆xt 0 εt
1
which incorporates the intercept into the cointegration vector. This is known as the “restricted
constant” case.
3. m = 0 and k = 0: This case is the most restrictive: clearly, neither xt nor yt are trended, and
the mean distance between them is zero. The vector µ0 is also 0, which explains why this case
is referred to as “no constant.”
In most cases, the choice between these three possibilities is based on a mix of empirical obser-
vation and economic reasoning. If the variables under consideration seem to follow a linear trend
Chapter 30. Cointegration and Vector Error Correction Models 273
then we should not place any restriction on the intercept. Otherwise, the question arises of whether
it makes sense to specify a cointegration relationship which includes a non-zero intercept. One ex-
ample where this is appropriate is the relationship between two interest rates: generally these are
not trended, but the VAR might still have an intercept because the difference between the two (the
“interest rate spread”) might be stationary around a non-zero mean (for example, because of a risk
or liquidity premium).
The previous example can be generalized in three directions:
1. If a VAR of order greater than 1 is considered, the algebra gets more convoluted but the
conclusions are identical.
2. If the VAR includes more than two endogenous variables the cointegration rank r can be
greater than 1. In this case, α is a matrix with r columns, and the case with restricted constant
entails the restriction that µ0 should be some linear combination of the columns of α.
3. If a linear trend is included in the model, the deterministic part of the VAR becomes µ0 + µ1 t.
The reasoning is practically the same as above except that the focus now centers on µ1 rather
than µ0 . The counterpart to the “restricted constant” case discussed above is a “restricted
trend” case, such that the cointegration relationships include a trend but the first differences
of the variables in question do not. In the case of an unrestricted trend, the trend appears
in both the cointegration relationships and the first differences, which corresponds to the
presence of a quadratic trend in the variables themselves (in levels).
In order to accommodate the five cases, gretl provides the following options to the coint2 and
vecm commands:
Note that for this command the above options are mutually exclusive. In addition, you have the
option of using the --seasonals options, for augmenting µt with centered seasonal dummies. In
each case, p-values are computed via the approximations devised by Doornik (1998).
open denmark
coint2 2 LRM LRY IBO IDE --rc --seasonals
Chapter 30. Cointegration and Vector Error Correction Models 274
In this case, the vector yt in equation (30.2) comprises the four variables LRM, LRY, IBO, IDE. The
number of lags equals p in (30.2) (that is, the number of lags of the model written in VAR form).
Part of the output is reported below:
Johansen test:
Number of equations = 4
Lag order = 2
Estimation period: 1974:3 - 1987:3 (T = 53)
Both the trace and λ-max tests accept the null hypothesis that the smallest eigenvalue is 0 (see the
last row of the table), so we may conclude that the series are in fact non-stationary. However, some
linear combination may be I(0), since the λ-max test rejects the hypothesis that the rank of Π is 0
(though the trace test gives less clear-cut evidence for this, with a p-value of 0.1284).
The coefficients that gretl produces are β̂, with B known as the matrix of unrestricted coefficients.
In terms of the underlying equilibrium relationship, the Phillips normalization expresses the system
1 For comparison with other studies, you may wish to normalize β differently. Using the set command you
can do set vecm_norm diag to select a normalization that simply scales the columns of the original β such that
βij = 1 for i = j and i ≤ r , as used in the empirical section of Boswijk and Doornik (2004). Another alternative is
set vecm_norm first, which scales β such that the elements on the first row equal 1. To suppress normalization
altogether, use set vecm_norm none. (To return to the default: set vecm_norm phillips.)
Chapter 30. Cointegration and Vector Error Correction Models 275
of r equilibrium relations as
open money.gdt
smpl 1954:1 1994:4
vecm 6 2 m infl cpr y tbr --rc
m 1.0000 0.0000
(0.0000) (0.0000)
infl 0.0000 1.0000
(0.0000) (0.0000)
cpr 0.56108 -24.367
(0.10638) (4.2113)
y -0.40446 -0.91166
(0.10277) (4.0683)
tbr -0.54293 24.786
(0.10962) (4.3394)
const -3.7483 16.751
(0.78082) (30.909)
Interpretation of the coefficients of the cointegration matrix β would be easier if a meaning could
be attached to each of its columns. This is possible by hypothesizing the existence of two long-run
relationships: a money demand equation
m = c1 + β1 infl + β2 y + β3 tbr
2 This data set is available in the verbeek data package; see http://gretl.sourceforge.net/gretl_data.html.
Chapter 30. Cointegration and Vector Error Correction Models 276
This renormalization can be accomplished by means of the restrict command, to be given after
the vecm command or, in the graphical interface, by selecting the “Test, Linear Restrictions” menu
entry. The syntax for entering the restrictions should be fairly obvious:3
restrict
b[1,1] = -1
b[1,3] = 0
b[2,1] = 0
b[2,3] = -1
end restrict
which produces
m -1.0000 0.0000
(0.0000) (0.0000)
infl -0.023026 0.041039
(0.0054666) (0.027790)
cpr 0.0000 -1.0000
(0.0000) (0.0000)
y 0.42545 -0.037414
(0.033718) (0.17140)
tbr -0.027790 1.0172
(0.0045445) (0.023102)
const 3.3625 0.68744
(0.25318) (1.2870)
Rb vec(β) = q (30.5)
and/or
Ra vec(α) = 0 (30.6)
Note that the β restriction may be non-homogeneous (q ≠ 0) but the α restriction must be homoge-
neous. Nonlinear restrictions are not supported, and neither are restrictions that cross between β
3 Note that in this context we are bending the usual matrix indexation convention, using the leading index to refer to
the column of β (the particular cointegrating vector). This is standard practice in the literature, and defensible insofar as
it is the columns of β (the cointegrating relations or equilibrium errors) that are of primary interest.
Chapter 30. Cointegration and Vector Error Correction Models 277
and α. When r > 1, such restrictions may be in common across all the columns of β (or α) or may
be specific to certain columns of these matrices. For useful discussions of this point see Boswijk
(1995) and Boswijk and Doornik (2004), section 4.4.
The restrictions (30.5) and (30.6) may be written in explicit form as
vec(β) = Hφ + h0 (30.7)
and
vec(α0 ) = Gψ (30.8)
respectively, where φ and ψ are the free parameter vectors associated with β and α respectively.
We may refer to the free parameters collectively as θ (the column vector formed by concatenating
φ and ψ). Gretl uses this representation internally when testing the restrictions.
If the list of restrictions that is passed to the restrict command contains more constraints than
necessary to achieve identification, then an LR test is performed. In addition, the restrict com-
mand can be given the --full switch, in which case full estimates for the restricted system are
printed (including the Γi terms) and the system thus restricted becomes the “current model” for
the purposes of further tests. Thus you are able to carry out cumulative tests, as in Chapter 7 of
Johansen (1995).
Syntax
The full syntax for specifying the restriction is an extension of that exemplified in the previous
section. Inside a restrict. . . end restrict block, valid statements are of the form
b[1,1] = 1.618
b[1,4] + 2*b[2,5] = 0
a[1,3] = 0
a[1,1] - a[1,2] = 0
Special syntax is used when a certain constraint should be applied to all columns of β: in this case,
one index is given for each b term, and the square brackets are dropped. Hence, the following
syntax
restrict
b1 + b2 = 0
end restrict
corresponds to
β11 β21
−β11 −β21
β=
β
13 β23
β14 β24
The same convention is used for α: when only one index is given for an “a” term the restriction is
presumed to apply to all r columns of α, or in other words the variable associated with the given
row of α is weakly exogenous. For instance, the formulation
Chapter 30. Cointegration and Vector Error Correction Models 278
restrict
a3 = 0
a4 = 0
end restrict
specifies that variables 3 and 4 do not respond to the deviation from equilibrium in the previous
period.4
A variant on the single-index syntax for common restrictions on α and β is available: you can
replace the index number with the name of the corresponding variable, in square brackets. For
example, instead of a3 = 0 one could write a[cpr] = 0, if the third variable in the system is
named cpr.
Finally, a shortcut (or anyway an alternative) is available for setting up complex restrictions (but
currently only in relation to β): you can specify Rb and q, as in Rb vec(β) = q, by giving the names
of previously defined matrices. For example,
matrix I4 = I(4)
matrix vR = I4**(I4~zeros(4,1))
matrix vq = mshape(I4,16,1)
restrict
R = vR
q = vq
end restrict
which manually imposes Phillips normalization on the β estimates for a system with cointegrating
rank 4.
There are two points to note in relation to this option. First, vec(β) is taken to include the coeffi-
cients on all terms within the cointegration space, including the restricted constant or trend, if any,
as well as any restricted exogenous variables. Second, it is acceptable to give an R matrix with a
number of columns equal to the number of rows of β; this variant is taken to specify a restriction
that is in common across all the columns of β.
An example
Brand and Cassola (2004) propose a money demand system for the Euro area, in which they postu-
late three long-run equilibrium relationships:
money demand m = βl l + βy y
Fisher equation π = φl
Expectation theory of l=s
interest rates
where m is real money demand, l and s are long- and short-term interest rates, y is output and
π is inflation.5 (The names for these variables in the gretl data file are m_p, rl, rs, y and infl,
respectively.)
The cointegration rank assumed by the authors is 3 and there are 5 variables, giving 15 elements
in the β matrix. 3 × 3 = 9 restrictions are required for identification, and a just-identified system
would have 15 − 9 = 6 free parameters. However, the postulated long-run relationships feature
only three free parameters, so the over-identification rank is 3.
4 Note that when two indices are given in a restriction on α the indexation is consistent with that for β restrictions:
the leading index denotes the cointegrating vector and the trailing index the equation number.
5 A traditional formulation of the Fisher equation would reverse the roles of the variables in the second equation,
but this detail is immaterial in the present context; moreover, the expectation theory of interest rates implies that the
third equilibrium relationship should include a constant for the liquidity premium. However, since in this example the
system is estimated with the constant term unrestricted, the liquidity premium gets absorbed into the system intercept
and disappears from zt .
Chapter 30. Cointegration and Vector Error Correction Models 279
restrict --full
b[1,1] = 1
b[1,2] = 0
b[1,4] = 0
b[2,1] = 0
b[2,2] = 1
b[2,4] = 0
b[2,5] = 0
b[3,1] = 0
b[3,2] = 0
b[3,3] = 1
b[3,4] = -1
b[3,5] = 0
end restrict
ll1 = $rlnl
Partial output:
Unrestricted loglikelihood (lu) = 116.60268
Restricted loglikelihood (lr) = 115.86451
2 * (lu - lr) = 1.47635
P(Chi-Square(3) > 1.47635) = 0.68774
Listing 30.1 replicates Table 4 on page 824 of the Brand and Cassola article.6 Note that we use
the $lnl accessor after the vecm command to store the unrestricted log-likelihood and the $rlnl
accessor after restrict for its restricted counterpart.
The example continues in script 30.2, where we perform further testing to check whether (a) the
income elasticity in the money demand equation is 1 (βy = 1) and (b) the Fisher relation is homo-
geneous (φ = 1). Since the --full switch was given to the initial restrict command, additional
restrictions can be applied without having to repeat the previous ones. (The second script contains
a few printf commands, which are not strictly necessary, to format the output nicely.) It turns out
that both of the additional hypotheses are rejected by the data, with p-values of 0.002 and 0.004.
restrict
b[2,3] = -1
end restrict
ll_hfh = $rlnl
Output:
Testing zero restrictions in cointegration space:
LR-test, rank = 3: chi^2(3) = 1.4763 [0.6877]
Unit income elasticity: LR-test, rank = 3:
chi^2(4) = 17.2071 [0.0018]
Homogeneity in the Fisher hypothesis:
LR-test, rank = 3: chi^2(4) = 15.547 [0.0037]
Another type of test that is commonly performed is the “weak exogeneity” test. In this context, a
variable is said to be weakly exogenous if all coefficients on the corresponding row in the α matrix
are zero. If this is the case, that variable does not adjust to deviations from any of the long-run
equilibria and can be considered an autonomous driving force of the whole system.
The code in Listing 30.3 performs this test for each variable in turn, thus replicating the first
column of Table 6 on page 825 of Brand and Cassola (2004). The results show that weak exogeneity
might perhaps be accepted for the long-term interest rate and real GDP (p-values 0.07 and 0.08
respectively).
restrict
a2 = 0
end restrict
ts_p = 2*(ll0 - $rlnl)
restrict
a3 = 0
end restrict
ts_l = 2*(ll0 - $rlnl)
restrict
a4 = 0
end restrict
ts_s = 2*(ll0 - $rlnl)
restrict
a5 = 0
end restrict
ts_y = 2*(ll0 - $rlnl)
A sufficient condition for identification is that the rank of J(θ) equals the number of free param-
eters. The rank of this matrix is evaluated by examination of its singular values at a randomly
selected point in the parameter space. For practical purposes we treat this condition as if it were
both necessary and sufficient; that is, we disregard the special cases where identification could be
achieved without this condition being met.7
We’d like to thank Luca Fanelli of the University of Bologna and Sven Schreiber of Goethe University Frankfurt for their
help in devising torture-tests for gretl’s VECM code.
Chapter 30. Cointegration and Vector Error Correction Models 283
optimizer may end up at a local maximum (or, in the case of the switching algorithm, at a saddle
point).
The solution (or lack thereof) may be sensitive to the initial value selected for θ. By default, gretl
selects a starting point using a deterministic method based on Boswijk (1995), but two further
options are available: the initialization may be adjusted using simulated annealing, or the user may
supply an explicit initial value for θ.
The default initialization method is:
3. vec(β0 ) = Hφ0 + h0 .
5. If α is restricted by vec(α0 ) = Gψ, then ψ0 = (G0 G)−1 G0 vec(α̂0 ) and vec(α00 ) = Gψ0 .
The details depend on whether the switching algorithm or LBFGS is used. For the switching algo-
rithm, there are two options for specifying the initial values. The more user-friendly one (for most
people, we suppose) is to specify a matrix that contains vec(β) followed by vec(α). For example:
Chapter 30. Cointegration and Vector Error Correction Models 284
open denmark.gdt
vecm 2 1 LRM LRY IBO IDE --rc --seasonals
In this example—from Johansen (1995) — the cointegration rank is 1 and there are 4 variables.
However, the model includes a restricted constant (the --rc flag) so that β has 5 elements. The α
matrix has 4 elements, one per equation. So the matrix BA may be read as
(β1 , β2 , β3 , β4 , β5 , α1 , α2 , α3 , α4 )
The other option, which is compulsory when using LBFGS, is to specify the initial values in terms
of the free parameters, φ and ψ. Getting this right is somewhat less obvious. As mentioned above,
the implicit-form restriction Rvec(β) = q has explicit form vec(β) = Hφ + h0 , where H = R⊥ , the
right nullspace of R. The vector φ is shorter, by the number of restrictions, than vec(β). The
savvy user will then see what needs to be done. The other point to take into account is that if α is
unrestricted, the effective length of ψ is 0, since it is then optimal to compute α using Johansen’s
formula, conditional on β (equation 30.11 above). The example above could be rewritten as:
open denmark.gdt
vecm 2 1 LRM LRY IBO IDE --rc --seasonals
In this more economical formulation the initializer specifies only the two free parameters in φ (5
elements in β minus 3 restrictions). There is no call to give values for ψ since α is unrestricted.
Scale removal
Consider a simpler version of the restriction discussed in the previous section, namely,
restrict
b[1] = 1
b[1] + b[2] = 0
end restrict
This restriction comprises a substantive, testable requirement — that β1 and β2 sum to zero — and
a normalization or scaling, β1 = 1. The question arises, might it be easier and more reliable to
maximize the likelihood without imposing β1 = 1?10 If so, we could record this normalization,
remove it for the purpose of maximizing the likelihood, then reimpose it by scaling the result.
Unfortunately it is not possible to say in advance whether “scale removal” of this sort will give
better results for any particular estimation problem. However, this does seem to be the case more
often than not. Gretl therefore performs scale removal where feasible, unless you
10 As a numerical matter, that is. In principle this should make no difference.
Chapter 30. Cointegration and Vector Error Correction Models 285
• explicitly forbid this, by giving the --no-scaling option flag to the restrict command; or
Scale removal is deemed infeasible if there are any cross-column restrictions on β, or any non-
homogeneous restrictions involving more than one element of β.
In addition, experimentation has suggested to us that scale removal is inadvisable if the system is
just identified with the normalization(s) included, so we do not do it in that case. By “just identified”
we mean that the system would not be identified if any of the restrictions were removed. On that
criterion the above example is not just identified, since the removal of the second restriction would
not affect identification; and gretl would in fact perform scale removal in this case unless the user
specified otherwise.
Chapter 31
Multivariate models
By a multivariate model we mean one that includes more than one dependent variable. Certain
specific types of multivariate model for time-series data are discussed elsewhere: chapter 29 deals
with VARs and chapter 30 with VECMs. Here we discuss two general sorts of multivariate model,
implemented in gretl via the system command: SUR systems (Seemingly Unrelated Regressions),
in which all the regressors are taken to be exogenous and interest centers on the covariance of the
error term across equations; and simultaneous systems, in which some regressors are assumed to
be endogenous.
In this chapter we give an account of the syntax and use of the system command and its compan-
ions, restrict and estimate; we also explain the options and accessors available in connection
with multivariate models.
• An estimation method is specified for the system. This is done by following system with
an expression of the form method=estimator, where estimator must be one of ols (Ordinary
Least Squares), tsls (Two-Stage Least Squares), sur (Seemingly Unrelated Regressions), 3sls
(Three-Stage Least Squares), liml (Limited Information Maximum Likelihood) or fiml (Full
Information Maximum Likelihood). Two examples:
system method=sur
system method=fiml
OLS, TSLS and LIML are, of course, single-equation methods rather than true system estima-
tors; they are included to facilitate comparisons.
• The system is assigned a name. This is done by giving the name first, followed by a back-
arrow, “<-”, followed by system. If the name contains spaces it must be enclosed in double-
quotes. Here are two examples:
Note, however, that this naming method is not available within a user-defined function, only
in the main body of a gretl script.
If the initial system line is augmented in the first way, the effect is that the system is estimated as
soon as its definition is completed, using the specified method. The effect of the second option is
286
Chapter 31. Multivariate models 287
that the system can then be referenced by the assigned name for the purposes of the restrict and
estimate commands; in the gretl GUI an additional effect is that an icon for the system is added
to the “Session view”.
These two possibilities can be combined, as in
In this example the system is estimated immediately via Three-Stage Least Squares, and is also
available for subsequent use under the name mysys.
If the system is not named via the back-arrow mechanism, it is still available for subsequent use
via restrict and estimate; in this case you should use the generic name $system to refer to the
last-defined multivariate system.
system method=sur
equation y1 const x1
equation y2 const x2
end system
For simultaneous systems it is necessary to determine which regressors are endogenous and which
exogenous. By default all regressors are treated as exogenous, except that any variable that appears
as the dependent variable in one equation is automatically treated as endogeous if it appears as a
regressor elsewhere. However, an explicit list of endogenous regressors may be supplied follow-
ing the equations lines: this takes the form of the keyword endog followed by the names or ID
numbers of the relevant regressors.
When estimation is via TSLS or 3SLS it is possible to specify a particular set of instruments for each
equation. This is done by giving the equation lists in the format used with the tsls command:
first the dependent variable, then the regressors, then a semicolon followed by the instruments, as
in
system method=3sls
equation y1 const x11 x12 ; const x11 z1
equation y2 const x21 x22 ; const x21 z2
end system
An alternative way of specifying instruments is to insert an extra line starting with instr, followed
by the list of variables acting as instruments. This is especially useful for specifying the system with
the equations keyword; see the following subsection. As in tsls, any regressors that are not also
listed as instruments are treated as endogenous, so in the example above x11 and x21 are treated
as exogenous while x21 and x22 are endogenous, and instrumented by z1 and z2 respectively.
One more sort of statement is allowed in a system block: that is, the keyword identity followed by
an equation that defines an accounting relationship, rather then a stochastic one, between variables.
For example,
identity Y = C + I + G + X
Chapter 31. Multivariate models 288
There can be more than one identity in a system block. But note that these statements are specific
to estimation via FIML; they are ignored for other estimators.
The first case is straightforward; the second requires a little more explanation. Suppose we have a
two-equation system with regressors given by the lists xlist1 and xlist2. We can then define a
suitable array as follows:
open denmark
list LHS = LRM LRY
list RHS1 = const LRM(-1) IBO(-1) IDE(-1)
list RHS2 = const LRY(-1) IBO(-1)
lists RHS = defarray(RHS1, RHS2)
system method=ols
equations LHS RHS
end system
As mentioned above, the option of assigning a specific name to a system is not available within
functions, but the generic identifier $system can be used to similar effect. The following example
illustrates how one can define a system, estimate it via two methods, apply a restriction, then
re-estimate it subject to the restriction.
restrict sysname
end restrict
As illustrated above, you can use the method tag to specify an estimation method with the estimate
command. If the system has already been estimated you can omit this tag and the previous method
is used again.
The estimate command is the main locus for options regarding the details of estimation. The
available options are as follows:
• If the estimation method is SUR or 3SLS and the --iterate flag is given, the estimator will be
iterated. In the case of SUR, if the procedure converges the results are maximum likelihood
estimates. Iteration of three-stage least squares, however, does not in general converge on the
full-information maximum likelihood results. This flag is ignored for other estimators.
• If the equation-by-equation estimators OLS or TSLS are chosen, the default is to apply a de-
grees of freedom correction when calculating standard errors. This can be suppressed using
the --no-df-corr flag. This flag has no effect with the other estimators; no degrees of free-
dom correction is applied in any case.
• By default, the formula used in calculating the elements of the cross-equation covariance
matrix is
û0 ûj
σ̂ij = i
T
where T is the sample size and ûi is the vector of residuals from equation i. But if the
--geomean flag is given, a degrees of freedom correction is applied: the formula is
û0i ûj
σ̂ij = q
(T − ki )(T − kj )
• If an iterative method is specified, the --verbose option calls for printing of the details of
the iterations.
• When the system estimator is SUR or 3SLS the cross-equation covariance matrix is initially
estimated via OLS or TSLS, respectively. In the case of a system subject to restrictions the
question arises: should the initial single-equation estimator be restricted or unrestricted?
The default is the former, but the --unrestrict-init flag can be used to select unrestricted
initialization. (Note that this is unlikely to make much difference if the --iterate option is
given.)
Chapter 31. Multivariate models 290
where σ̂i2 is û0i ûi /T from the individual OLS regressions. In both cases the test statistic is dis-
tributed asymptotically as χ 2 with g(g − 1)/2 degrees of freedom.
where yt represents the vector of endogenous variables in period t, xt denotes the vector of ex-
ogenous variables, and p is the maximum lag of the endogenous regressors. The structural-form
matrices can be retrieved as $sysGamma, $sysA and $sysB respectively. If yt is m × 1 and xt is
n × 1, then Γ is m × m and B is m × n. If the system contains no lags of the endogenous variables
then the A matrix is not defined, otherwise A is the horizontal concatenation of A1 , . . . , Ap , and is
therefore m × mp.
From the structural form it is straightforward to obtain the reduced form, namely,
p
X
yt = Γ −1 Ai yt−i + Γ −1 Bxt + vt
i=1
where vt ≡ Γ −1 t . The reduced form is used by gretl to generate forecasts in response to the fcast
command. This means that—in contrast to single-equation estimation — the values produced via
fcast for a static, within-sample forecast will in general differ from the fitted values retrieved via
$yhat. The fitted values for equation i represent the expectation of yti conditional on the contem-
poraneous values of all the regressors, while the fcast values are conditional on the exogenous
and predetermined variables only.
The above account has to be qualified for the case where a system is set up for estimation via TSLS
or 3SLS using a specific list of instruments per equation, as described in section 31.1. In that case
it is possible to include more endogenous regressors than explicit equations (although, of course,
there must be sufficient instruments to achieve identification). In such systems endogenous re-
gressors that have no associated explicit equation are treated “as if” exogenous when constructing
the structural-form matrices. This means that forecasts are conditional on the observed values of
the “extra” endogenous regressors rather than solely on the values of the exogenous and predeter-
mined variables.
Chapter 32
Forecasting
32.1 Introduction
In some econometric contexts forecasting is the prime objective: one wants estimates of the future
values of certain variables to reduce the uncertainty attaching to current decision making. In other
contexts where real-time forecasting is not the focus prediction may nonetheless be an important
moment in the analysis. For example, out-of-sample prediction can provide a useful check on
the validity of an econometric model. In other cases we are interested in questions of “what if”:
for example, how might macroeconomic outcomes have differed over a certain period if a different
policy had been pursued? In the latter cases “prediction” need not be a matter of actually projecting
into the future but in any case it involves generating fitted values from a given model. The term
“postdiction” might be more accurate but it is not commonly used; we tend to talk of prediction
even when there is no true forecast in view.
This chapter offers an overview of the methods available within gretl for forecasting or prediction
(whether forward in time or not) and explicates some of the finer points of the relevant commands.
series yh = $yhat
If the model in question takes the form of a system of equations, $yhat returns a matrix, each
column of which contains the fitted values for a particular dependent variable. To extract the fitted
series for, e.g., the dependent variable in the second equation, do
matrix Yh = $yhat
series yh2 = Yh[,2]
Having obtained a series of fitted values, you can use the fcstats function to produce a vector of
statistics that characterize the accuracy of the predictions (see section 32.4 below).
The gretl GUI offers several ways of accessing and examining within-sample predictions. In the
model display window the Save menu contains an item for saving fitted values, the Graphs menu
allows plotting of fitted versus actual values, and the Analysis menu offers a display of actual, fitted
and residual values.
291
Chapter 32. Forecasting 292
• Use the smpl command to adjust the sample range prior to invoking fcast.
• Use the optional startobs and endobs arguments to fcast (which should come right after the
command word). These values set the forecast range independently of the sample range.
What if one wants to generate a true forecast that goes beyond the available data? In that case
one can use the dataset command with the addobs parameter to add extra observations before
forecasting. For example:
But this will work as stated only if the set of regressors in xlist does not contain any stochastic
regressors other than lags of y. The dataset addobs command attempts to detect and extrapolate
certain common deterministic variables (e.g., time trend, periodic dummy variables). In addition,
lagged values of the dependent variable can be supported via a dynamic forecast (see below for
discussion of the static/dynamic distinction). But “future” values of any other included regressors
must be supplied before such a forecast is possible. Note that specific values in a series can be
set directly by date, for example: x1[2009:1] = 120.5. Or, if the assumption of no change in the
regressors is warranted, one can do something like this:
loop t=2009:1..2009:4
loop foreach i xlist
$i[t] = $i[2008:4]
endloop
endloop
yt = α0 + α1 yt−1 + t (32.1)
Chapter 32. Forecasting 293
In some cases the presence of a lagged dependent variable is implicit in the dynamics of the error
term, for example
yt = β + ut
ut = ρut−1 + t
Suppose we want to forecast y for period s using a dynamic model, say (32.1) for example. If
we have data on y available for period s − 1 we could form a fitted value in the usual way: ŷs =
α̂0 + α̂1 ys−1 . But suppose that data are available only up to s − 2. In that case we can apply the
chain rule of forecasting:
This is what is called a dynamic forecast. A static forecast, on the other hand, is simply a fitted
value (even if it happens to be computed out-of-sample).
T T
1 X et 1 X e
t
MPE = 100 MAPE = 100
T t=1 yt T t=1 yt
A further relevant statistic is Theil’s U (Theil, 1966), defined as the positive square root of
T −1
!2 T −1 !2 −1
2 1 X ft+1 − yt+1 1 X yt+1 − yt
U = ·
T t=1 yt T t=1 yt
The more accurate the forecasts, the lower the value of Theil’s U , which has a minimum of 0.1
This measure can be interpreted as the ratio of the RMSE of the proposed forecasting model to the
RMSE of a naïve model which simply predicts yt+1 = yt for all t. The naïve model yields U = 1;
values less than 1 indicate an improvement relative to this benchmark and values greater than 1 a
deterioration.
1 This statistic is sometimes called U , to distinguish it from a related but different U defined in an earlier work by
2
Theil (1961). It seems to be generally accepted that the later version of Theil’s U is a superior statistic, so we ignore the
earlier version here.
Chapter 32. Forecasting 294
In addition, Theil (1966, pp. 33–36) proposed a decomposition of the MSE which can be useful in
evaluating a set of forecasts. He showed that the MSE could be broken down into three non-negative
components as follows
2 2
MSE = f¯ − ȳ + sf − r sy + 1 − r 2 s 2 y
where f¯ and ȳ are the sample means of the forecasts and the observations, sf and sy are the re-
spective standard deviations (using T in the denominator), and r is the sample correlation between
y and f . Dividing through by MSE we get
2 2
f¯ − ȳ sf − r sy 1 − r 2 sy2
+ + =1 (32.2)
MSE MSE MSE
Theil labeled the three terms on the left-hand side of (32.2) the bias proportion (U M ), regression
proportion (U R ) and disturbance proportion (U D ), respectively. If y and f represent the in-sample
observations of the dependent variable and the fitted values from a linear regression then the first
two components, U M and U R , will be zero (apart from rounding error), and the entire MSE will be
accounted for by the unsystematic part, U D . In the case of out-of-sample prediction, however (or
“prediction” over a sub-sample of the data used in the regression), U M and U R are not necessarily
close to zero, although this is a desirable property for a forecast to have. U M differs from zero if
and only if the mean of the forecasts differs from the mean of the realizations, and U R is non-zero
if and only if the slope of a simple regression of the realizations on the forecasts differs from 1.
The above-mentioned statistics are printed as part of the output of the fcast command. They can
also be retrieved in the form of a column vector using the function fcstats, which takes two series
arguments corresponding to y and f . The vector returned is
0
ME RMSE MAE MPE MAPE U UM UR UD
(Note that the MSE is not included since it can easily be obtained given the RMSE.) The series given
as arguments to fcstats must not contain any missing values in the currently defined sample
range; use the smpl command to adjust the range if needed.
yields
33.1 Introduction
The original interface to the treatment of linear state space models in gretl (which has now been
retired) predated the development of gretl’s “bundle” type.1 In 2016 it struck the gretl authors that
the bundle type could be used to advantage in reformulating the user interface; we suspect people
will find the new interface substantially easier to use. The new interface makes it more practical
to split long portions of code into smaller, dedicated functions, which are arguably easier to debug
and maintain; hopefully this may lead to more function packages using gretl’s native state space
facilities. Moreover, the new design is such that certain operations that would have required a
good deal of low-level coding under the old regime are now simple, fast and natural — for example,
“disturbance smoothing” (see section 33.7) or setting up time-varying system matrices (see section
33.9).
Technically, a gretl bundle is an associative array; that is, an all-purpose container into which you
can put matrices, scalars, strings and so on, identified by keys. A Kalman bundle is a somewhat
regimented version of the same. There are two sets of “reserved” keys, one set pertaining to input
matrices and the other to results generated by the Kalman functions. Beyond that, however, you
are free to add to the bundle whatever extra elements you think may be useful.
Here are the main features of the new interface “at a glance”:
• The Kalman structure, which used to be entirely sui generis, is implemented as a bundle.
• You can therefore have as many named models as you like in existence at one time, and you
can pass them as arguments to functions.
• The “block” syntax for defining a filter is replaced by the function ksetup.
• The signatures of the functions kfilter and ksmooth are much simplified. For the most
part you retrieve results by reading bundle members instead of poking in matrix-pointer ar-
guments.
• The mechanism for handling time-varying system matrices has been simplified and general-
ized.
33.2 Notation
In this document our basic representation of a state-space model is given by the following pair of
equations:
αt+1 = Ft αt + vt (33.1)
yt = A0t xt + Ht0 αt + wt (33.2)
1 If anyone needs documentation for the original interface, available in gretl 2016d and earlier versions, it can be
found at http://gretl.sourceforge.net/papers/kalman_old.pdf.
297
Chapter 33. State Space Modeling 298
where (33.1) is the state transition equation and (33.2) is the observation or measurement equation.
The state vector, αt , is (r × 1) and the vector of observables, yt , is (n × 1); xt is a (k × 1) vector of
exogenous variables. The (r × 1) vector vt and the (n × 1) vector wt are assumed to be vector white
noise:
y T ×n obsy
H r ×n obsymat
F r ×r statemat
Q r ×r statevar
For example,
The names of these input matrices don’t matter; in fact they may be anonymous matrices con-
structed on the fly. But if and when you wish to copy them out of the bundle, you must use the
specified keys:
matrix H = SSmod.obsymat
Although all the arguments are in concept matrices, as a convenience you may give obsy as a series
or list of series, and the other arguments can be given as scalars if in context they are 1 × 1.
If applicable you may specify any of the following optional input matrices:3
However, these optional matrices are not passed to ksetup, rather you add them to the bundle
returned by ksetup (under their reserved keys) as you usually add elements to a bundle:
2 An optional, fifth argument is used for initializing a state space model in which there is some non-zero correlation
SSmod.obsxmat = A
Once setup is complete the dimensions of the model, r , n, k and T , become available as scalar
members of the bundle (under their own names).
It might appear that the obsx (x) and obsxmat (A) matrices must go together — either both are
given or neither is given. But an exception is granted for convenience. If the observation equation
includes a constant but no additional exogenous variables, you can give a (1×n) value for A without
having to specify obsx. More generally, if the row dimension of A is 1 greater than the column
dimension of x, it is assumed that the first element of A is associated with an implicit column of
ones.
The automatic initialization of P1|0 (in case inivar is not specified) is applicable only if all the
eigenvalues of F lie inside the unit circle. In this case, the variance for the marginal distribution of
αt is well defined and equals
which is the formula used internally. Otherwise, we apply a diffuse prior, setting P1|0 = κIr with
κ = 107 . You can impose this diffuse prior from the outset by setting a non-zero scalar value on
the bundle under the (reserved) key diffuse:
SSmod.diffuse = 1
• You can replace the coefficient matrices obsymat, statemat, statevar and (if applicable)
obsvar and/or obsxmat in a given bundle if you wish — but only on condition that the re-
placement matrix has the same dimensions as the original. Or in other words, the dimensions
r , n and k are set once and for all by the initialization (section 33.3).
• You can replace the data matrices obsy and (if applicable) obsx subject to the condition that
the number of columns is unchanged (n and k respectively): the time-series length, T , is
mutable. But note that if obsx is present it must have at least as many rows as obsy (which
defines T for the bundle).
• None of the input matrices just mentioned can be deleted from the bundle.
• Output matrices that are automatically added to the bundle by the functions described in the
following sections can be deleted (if you don’t need them and want to save on storage). But
they cannot be replaced by arbitrary user content.
• The only other “special” member that can be deleted is the function call (string) that is dis-
cussed in section 33.9.
Nonetheless, in the “user area” of the bundle (that is, under keys other than the reserved ones
noted in this chapter), the usual rules apply.
For all the “k” functions described below the first argument (in most cases the only argument) must
be a pointer to a bundle obtained via ksetup. Any old bundle will not do. A “pointer to bundle”
is specified by prefixing the name of the bundle with an ampersand, as in “&SSmod”. Passing the
argument in pointer form allows these functions to modify the content of the bundle.
Chapter 33. State Space Modeling 300
(but see section 33.9 below for modifications to these formulae for the case of a diffuse prior). The
key llt gives access to a T -vector, element t of which is
1h i
`t = − n log(2π ) + log |Σt | + et0 Σ−1
t et
2
Five additional matrices also become available. Each of these has T rows, one for each time-step;
the contents of the rows are shown in the following listing.
1. Forecast errors for the observable variables, et0 , n columns: key prederr.
2. Variance matrix for the forecast errors, vech(Σt )0 , n(n + 1)/2 columns: key pevar.
0
3. Estimate of the state vector, α̂t|t−1 , r columns: key state.
4. MSE of estimate of the state vector, vech(Pt|t−1 )0 , r (r + 1)/2 columns: key stvar.
This matrix is rarely required by the user. However, it is the key quantity in the filtering algorithm,
so we make it available under a dedicated key for diagnostic purposes in case numerical problems
should arise. For example, the following retrieves the gain after a filtering operation:
kfilter(&SSmod)
matrix G = SSmod.gain
Then, for example, if you want to retrieve the matrix K at time 10, you need to reshape the tenth
row of the matrix G into the appropriate dimensions:
Note that since ksmooth starts with a forward pass, it can be run without a prior call to kfilter.
This may appear to be useless duplication, but in fact it enables an efficient scripting option. The
main utility of the forward pass lies in the calculation of the log-likelihood in the context of estima-
tion; however, if a state space model contains no parameters that have to be estimated, the model
setup can be followed directly by a call to ksmooth. (And the same goes for kdsmooth below.)
The backward-pass algorithm is as follows: for t = T , . . . , 1
Lt = Ft − Kt Ht0
ut−1 = Ht Σ−1 0
t et + Lt ut
Ut−1 = Ht Σ−1 0 0
t Ht + Lt Ut Lt
α̂t|T = α̂t|t−1 + Pt|t−1 ut−1
Pt|T = Pt|t−1 − Pt|t−1 Ut−1 Pt|t−1
(This assumes the observation equation has a stochastic component; if it does not, then smdist is
just T × r .)
An associated dispersion measure is provided under the key smdisterr. The precise definition
of this matrix depends on a second, optional Boolean parameter. Before describing the action of
this parameter we need to give a brief account of the two variance measures that are found in the
literature on disturbance smoothing. Our account runs in terms of the state disturbance, vt , but it
applies equally to the observation disturbance, wt , if present.
One measure of variance is the mean square distance of the inferred disturbances from zero (that
(1)
is, from their unconditional expectation). Let us call this Vt :
(1)
= E v̂t v̂t0
Vt
This measure is used in computing the so-called auxiliary residuals, which are advocated in Durbin
and Koopman (2012) as useful diagnostic tools. Auxiliary residuals for the state equation are ob-
(1)
tained by dividing v̂t by the square roots of the associated diagonal elements of Vt . In computing
this matrix we use the formulae given in Koopman et al. (1999, section 4.4).
A second measure of variance is the mean squared distance of the inferred disturbances around
(2)
their true values, or in other words the mean squared error, which we’ll write as Vt .
(2)
= E (v̂t − vt ) (v̂t − vt )0 | y1 , . . . , yT
Vt
4 See I. Karibzhanov’s clear exposition at http://karibzhanov.com/help/kalcvs.htm.
Chapter 33. State Space Modeling 302
We calculate this matrix using the formulae given in Durbin and Koopman (2012, section 4.5.2). Its
diagonal elements can be used to form confidence intervals for the true disturbances.
We are now ready to state what gretl provides under the key smdisterr. If the optional second
(2)
argument is present and non-zero the results are based on Vt , otherwise (that is, by default) they
(1)
are based on Vt . In either case row t of smdisterr contains the square roots of the diagonal
elements of the matrix in question: the first r elements pertain to the state disturbances and the
following n elements to the observation equation (if applicable). Like smdist, smdisterr has T
rows and either r + n or r columns depending on whether or not there’s a disturbance term in the
observation equation. We return standard deviations rather than variances since most of the time
it’s the former that users will actually want.
Section 33.10 presents a script which exercises the disturbance smoother and illustrates the differ-
(1) (2)
ence between Vt and Vt .
and post-multiplying U by Z 0 (although it’s not necessary to form Z explicitly if the disturbance
variance matrices are diagonal). This is not particularly onerous if Ω is time-invariant; gretl’s
psdroot function can be used to form Z if it’s needed. If Q and/or R are time-varying, however,
this scaling can become quite burdensome. As a convenience, the ancillary function ksimdata can
be used to pre-process the U matrix automatically. Here’s an example (we assume that a bundle
named B has been obtained via ksetup):
Time-variation or no, ksimdata should ensure that the disturbances are scaled to the variances
specified by Qt and Rt .
If the disturbances are cross-correlated (see section 33.9) then the matrix argument to ksimul
should be T × p, each row holding εt0 . In this case no prior scaling is required since it is assumed
that εt ∼ N(0, I) and the disturbances for the respective equations, Bt εt and Ct εt , are computed
internally, with regard to time-variation if applicable.
For the purposes of ksimul the time-series length, T , is defined by the number of rows of the
supplied disturbance matrix. This need not equal the original T value (set from obsy in the initial
call to ksetup), since obsy is ignored under simulation. However, if the model includes exogenous
Chapter 33. State Space Modeling 303
variables in the observation equation (obsx) and the simulation length is greater than the original
T , the simulation will run out of data and fail unless you supply a larger x matrix. This can be
done in either of two ways. You can add a suitably sized matrix to the Kalman bundle under the
key simx; if present, this will be used in preference to obsx for simulation. Or if you don’t mind
over-writing the original obsx you can substitute another matrix with the required number of rows
under that key.
By default, the value returned by ksimul is a (T × n) matrix holding simulated values for the vector
of observables at each time step; that is, row t holds ỹt0 , where the tilde indicates a simulated
quantity. To obtain a record of the simulated state, supply a non-zero value for the final, optional
argument. In that case the returned matrix is T × (r + n) and contains both the simulated state
and the simulated observable; row t holds (α̃t0 , ỹt0 ).
Note that the original obsy member of the bundle is not overwritten by ksimul, nor is state or
any other user-accessible output matrix. On exit from ksimul the prior T value is restored.
The recursion that yields ỹ and α̃ is as follows: for t = 1, . . . , T
ỹt = A0t xt + Ht0 α̃t + wt
α̃t+1 = Ft α̃t + vt
This implies that a value for α̃1 is required to get things started. You can add such a value to a
Kalman bundle under the reserved key simstart. If this member is not present in the bundle,
α̃1 defaults to the value given under the key inistate, or if that in turn is not present, to a zero
vector. A further note on initializing the state can be found in section 33.9.
In case you wish to retrieve or update information on the variance of the disturbances, note that in
the cross-correlated case the bundle keys statevar and obsvar are taken to designate the factors
B and C respectively.
The term µ is never strictly necessary: the system (33.1) and (33.2) without µ is general enough
to accommodate such a term, by absorbing it as an extra (non time-varying) element in the state
vector. However, this comes at the cost of expanding all the matrices that touch the state (α,
F , v, Q, H), making the model relatively awkward to formulate and forecasts more expensive to
compute. Therefore, we adopt the convention above on practical grounds.
The (r × 1) vector µ can be specified as a bundle member by the name stconst. Despite its name,
this matrix can be stipulated as time varying, as explained in the next section.
Time-varying matrices
Any or all of the matrices obsymat, obsxmat, obsvar, statemat, statevar and stconst may be
time-varying. In that case you must supply the name of a function to be called to update the matrix
or matrices in question: you add this to the bundle as a string, under the key timevar_call.5 For
example, if just obsymat (Ht ) should be updated, by a function named TV_H, you would write
SSmod.timevar_call = "TV_H"
The function that plays this role will be called at each time-step of the filtering or simulation
operation, prior to performing any calculations. It should have a single bundle-pointer parameter,
by means of which it will be passed a pointer to the Kalman bundle to which the call is attached. Its
return value, if any, will not be used, so generally it returns nothing (is of type void). However, you
can use gretl’s funcerr keyword to raise an error if things seem to be going wrong; see chapter 13
for details.
Besides the bundle members noted above, a time variation function has access to the current (1-
based) time step, under the reserved key t, and the n-vector containing the forecast error from the
previous time step, et−1 , under the key uhat; when t = 1 the latter will be a zero vector.
If any additional information is needed for performing the update it can be placed in the bundle
under a user-specified key. So, for example, a simple updater for a (1 × 1) H matrix might look like
this:
where in this case we assume that b.Hvals is T × r n, with row t holding the (transposed) vec of
Ht , and b.Fvals is T × r (r + 1)/2, with row t holding the vech of Ft . Simpler variants (e.g. just one
element of a relevant matrix in question is changed) and more complex variants — say, involving
some sort of conditionality—are also possible in this framework.
It is worth noting that this setup lends itself to a much wider scope than time-varying system
matrices. In fact, this syntax allows for the possibility of executing user-defined operations at each
step. The function that goes under timevar_call can read all the elements of the model bundle
and can modify several of them: the system matrices (which can therefore be made time-varying)
as well as the user-defined elements.
An extended example of use of the time-variation facility is presented in section 33.10.
5 The choice of the name for the function itself is of course totally up to the user.
Chapter 33. State Space Modeling 305
and
T
1 X
σ̂ 2 = e0 Σ−1 et
nT − d t=1 t t
α1 = a + Zv0
matrix Z = psdroot(B.inivar)
B.simstart = B.inistate + Z * mnormal(B.r, 1)
ARMA estimation
Functions illustrated in this example: ksetup, kfilter.
As is well known, the Kalman filter provides a very efficient way to compute the likelihood of ARMA
models; as an example, take an ARMA(1,1) model
yt = φyt−1 + εt + θεt−1
One of the ways the above equation can be cast in state-space form is by defining a latent process
ξt = (1 − φL)−1 εt . The observation equation corresponding to (33.2) is then
yt = ξt + θξt−1 (33.3)
Note that the observation equation (33.3) does not include an “error term”; this is equivalent to
saying that var(wt ) = 0 and there is therefore no need to add an obsvar element to the bundle.
Once the filter is set up, all it takes to compute the log-likelihood for given values of φ, θ and
σε2 is to execute the kfilter function and read the Kalman bundle’s lnl member (which holds
the total log-likelihood) or, more appropriately if the likelihood has to be maximized through mle,
llt, which gives the series of individual contribution to the log-likelihood for each observation. An
example is shown in script 33.1.
xt = µt + ε t
2
∆ µt = ηt ,
where is an observable time series xt and εt and ηt are mutually uncorrelated white noise processes;
the ratio between the two variances is the “smoothing” parameter: λ = VV (η (εt )
t)
. This is normally
calibrated rather than estimated, and a common choice for macro data is to make this a function
of the data periodicity: λ = 100 · p 2 , where p is the number of subperiods in a year. See Harvey
and Jaeger (1993) and King and Rebelo (1993) for details.
The most common usage of the HP filter is to retrieve an estimate of µt by using all the available
data, so that the estimate of the trend µt at some point t uses the future values of xt as well as data
points from the past. In some cases (for example, for forecasting purposes) this is undesirable, and
it may be preferable to construct µ̂t as a function of xt , xt−1 , . . . only. This point was famously
made by Stock and Watson (1999), and the resulting estimator is known as the one-sided HP filter.
A possible space-state representation of the model above is
" #
µt
xt = [1 0] + εt
µt−1
" # " #" #
µt 2 −1 µt−1
= + ηt .
µt−1 1 0 µt−2
and the estimates for µt can be easily obtained by running a forward filter for the one-sided version
and a smoothing pass for the two-sided one. The corresponding code implementing the filter
is shown in script 33.2, which also shows an example with the “housing starts” series from the
St. Louis Fed database. The example also compares the result of the function to that from the
native hpfilt() function.
Note that, in the case of the one-sided filter, a little trick is required in order to get the desired
result: the state matrix that the kfilter() function stores in the bundle is the estimate of α̂t|t−1 ,
whereas what is required is in fact α̂t|t . In order to circumvent the problem, we add an extra
observation to the end of the series and retrieve the one-step-ahead estimate of the lagged state.
Chapter 33. State Space Modeling 308
ssm = ksetup(my, H, F, Q)
ssm.obsvar = sqrt(lambda)
ssm.inistate = {2*y[1]-y[2] ; 3*y[1]-2*y[2]}
ssm.diffuse = 1
clear
open fedstl.bin
data houst
y = log(houst)
nulldata 200
set seed 101010
setobs 1 1 --special-time-series
/* ML estimation of variances */
mle ll = ERR ? NA : kb.llt
ERR = kfilter(&kb)
params kb.statevar kb.obsvar
end mle
The two unknown parameters σ12 and σ22 can be estimated via maximum likelihood. Script 33.3
provides an example of simulation and estimation of such a model. Since simulating the local level
model is trivial using ordinary gretl commands, we don’t use ksimul in this context.7
6 Note that the local level model, plus other common “Structural Time Series” models are implemented in the Struc-
Time-varying models
In order to illustrate state space models with time-varying system matrices, we will use time-varying
OLS. Suppose the DGP for an observable time series yt is given by
yt = β0 + β1,t xt + εt (33.4)
It is easy to see that the pair of equations above define a state space model, with equation (33.4)
as the measurement equation and (33.5) as the state transition equation. In other terms, β1,t is the
unobservable state, F = 1 and Q = V (ηt ) = ση2 ; as for the measurement equation, R = V (εt ) = σε2 ,
but the matrix multiplying β1,t , and hence H, is xt , which is clearly time-varying.
-0.05
-0.1
tvar_b1hat
-0.15
-0.2
-0.25
-0.3
1975 1980 1985 1990
Figure 33.1: Phillips Curve on Euro data: time-varying slope and 95% confidence interval
Once the system is framed as a state-space model, estimation of the three unknown parameters
β0 , σε2 and ση2 may proceed by maximum likelihood in a manner similar to example 33.1 and 33.3.
The sequence of slope coefficients β1,t is, instead, estimated by running the smoother, which also
yields a consistent estimate of the dispersion of the estimated state
Script 33.4 presents an example of the above, in which data from the AWM database are used to
estimate a Phillips Curve with time-varying slope:
where INFQ is a measure of quarterly inflation and URX a measure of unemployment. At the end
of the script, the evolution through time of the slope coefficient is plotted together with a 95%
confidence band—see Figure 33.1.
series. For convenience, the script is written as if equation (33.1) was modified into the equivalent formulation
αt = F αt−1 + vt ;
we tried to keep the script as simple as possible, so that the reader is free to focus on the interesting aspects.
Chapter 33. State Space Modeling 311
/* parameter initialization */
scalar b0 = mean(INFQ)
scalar s_obs = 0.1
scalar s_state = 0.1
/* bundle setup */
bundle B = ksetup(INFQ, 1, 1, 1)
matrix B.mX = {URX}
matrix B.depvar = {INFQ}
B.timevar_call = "at_each_step"
B.diffuse = 1
Disturbance smoothing
Functions illustrated in this example: ksetup, kdsmooth.
In section 33.7 we noted that the kdsmooth function can produce two different measures of the
dispersion of the smoothed disturbances, depending on the the value of the (optional) trailing
Boolean parameter. Here we show what these two measures are good for, using the famous Nile
flow data that have been much analysed in the state-space modeling literature. We focus on the
state equation; that is, the random-walk component of the observed series. For consistency with
the literature let us call the t-dated disturbance to the state ηt .
In the first call to kdsmooth we omit the optional switch and therefore compute E(η̂t η̂0t ) for each
t. This quantity is suitable for constructing the auxiliary residuals shown in the top panel of
Figure 33.2 (for similar plots see Koopman et al. (1999), Pelagatti (2011)). This plot suggests the
presence of a structural break shortly prior to 1900, as various authors have observed.
In the second kdsmooth call we ask gretl to compute instead E[(η̂t − ηt )(η̂t − ηt )0 |y1 , . . . , yT ], the
MSE of η̂t considered as an estimator of ηt . And in the lower panel of the Figure we plot η̂t along
with a 90% confidence band (roughly, ±1.64 times the RMSE). This reveals that, given the sampling
variance of η̂t , we’re not really sure that any of the ηt values were truly different from zero. The
resolution of the seeming conflict here is commonly reckoned to be that there was in fact a change
in mean around 1900, but besides that event there’s little evidence for a non-zero ση2 . Or in other
words the standard local level model is not really applicable to the data.
open nile.gdt
# ML variance estimates
scalar s2_eta = 1468.49
scalar s2_eps = 15099.7
kdsmooth(&LLM)
series eta_aux = LLM.smdist[,1] ./ LLM.smdisterr[,1]
series zero = 0
plot eta_aux
options time-series with-lines band=zero,const,2
option band-style=dash
literal set ylabel ’’
literal set title ’Auxiliary residual, state equation’
end plot --output=display
kdsmooth(&LLM, 1)
series etahat = LLM.smdist[,1]
series sdeta = LLM.smdisterr[,1]
plot etahat
options time-series with-lines band=etahat,sdeta,1.64485
option band-style=dash
literal set ylabel ’’
literal set title ’State disturbance with 90% confidence band’
end plot --output=display
Chapter 33. State Space Modeling 313
-1
-2
-3
-4
1880 1900 1920 1940 1960
100
80
60
40
20
-20
-40
-60
-80
-100
-120
1880 1900 1920 1940 1960
Figure 33.2: Nile data: auxiliary residuals and η̂t from disturbance smoother
Chapter 34
Numerical methods
Several functions are available to aid in the construction of special-purpose estimators: their pur-
pose is to find numerically approximate solutions to problems that in principle could be solved
analytically, but in practice cannot be, for one reason or another. In this chapter, we illustrate the
tools that gretl offers for optimization of functions, differentiation and integration.
BFGS
The BFGSmax function has two required arguments: a vector holding the initial values of a set of
parameters, and a call to a function that calculates the (scalar) criterion to be maximized, given
the current parameter values and any other relevant data. If the object is in fact minimization, this
function should return the negative of the criterion. On successful completion, BFGSmax returns the
maximized value of the criterion and the vector given via the first argument holds the parameter
values which produce the maximum. It is assumed here that the objective function is a user-defined
function (see Chapter 13) with the following general set-up:
The first argument contains the parameter vector (which should not be modified within the func-
tion) and the second may be used to hold “extra” values that are necessary to compute the objective
function, but are not the variables of the optimization problem. Here the pointer form is chosen
for the argument, but depending on the problem it could also be passed as a plain argument, with
our without the const modifier. For example, if the objective function were a log-likelihood, the
first argument would contain the parameters and the second one the data. Or, for more economic-
theory inclined readers, if the objective function were the utility of a consumer, the first argument
might contain the quantities of goods and the second one their prices and disposable income.
The operation of BFGS can be adjusted using the set variables bfgs_maxiter and bfgs_toler
(see Chapter 23). In addition you can provoke verbose output from the maximizer by assigning a
positive value to max_verbose, again via the set command.
The Rosenbrock function is often used as a test problem for optimization algorithms. It is also
known as “Rosenbrock’s Valley” or “Rosenbrock’s Banana Function”, on account of the fact that its
contour lines are banana-shaped. It is defined by:
f (x, y) = (1 − x)2 + 100(y − x 2 )2
The function has a global minimum at (x, y) = (1, 1) where f (x, y) = 0. Listing 34.1 shows a gretl
script that discovers the minimum using BFGSmax (giving a verbose account of progress). Note
314
Chapter 34. Numerical methods 315
that, in this particular case, the function to be maximized only depends on the parameters, so the
second parameter is omitted from the definition of the function Rosenbrock.
set lbfgs on
The primary case for using L-BFGS-B, however, is constrained optimization: this algorithm supports
constraints on the parameters in the form of minima and/or maxima. In gretl this is implemented
by the function BFGScmax (‘c’ for constrained). The syntax is basically similar to that of BFGSmax,
except that the first argument must followed by specification of a “bounds” matrix. This matrix
Chapter 34. Numerical methods 316
should have three columns and as many rows as there are constrained elements of the parameter
vector. Each row should hold the (1-based) index of the constrained parameter, followed by lower
and upper bounds. The values -$huge and $huge should be used to indicate that the parameter
is unconstrained downward or upward, respectively. For example, the following code constructs a
matrix to specify that the second element of the parameter vector must be non-negative, and the
fourth must lie between 0 and 1:
Newton–Raphson
BFGS, discussed above, is an excellent all-purpose maximizer, and about as robust as possible given
the limitations of digital computer arithmetic. The Newton–Raphson maximizer is not as robust,
but may converge much faster than BFGS for problems where the maximand is reasonably well
behaved— in particular, where it is anything like quadratic (see below). The case for using Newton–
Raphson is enhanced if it is possible to supply a function to calculate the Hessian analytically.
The gretl function NRmax, which implements the Newton–Raphson method, has a maximum of four
arguments. The first two (required) arguments are exactly as for BFGS: an initial parameter vector,
and a function call which returns the maximand given the parameters. The (optional) third argu-
ment is again as in BFGS: a function call that calculates the gradient. Specific to NRmax is an optional
fourth argument, namely a function call to calculate the (negative) Hessian. The first argument of
this function must be a pre-defined matrix of the right dimension to hold the Hessian — that is, a
k × k matrix, where k is the length of the parameter vector — given in “pointer” form. The second
argument should be the parameter vector (optionally in pointer form). Other data may be passed
as additional arguments as needed. Similarly to the case with the gradient, if the fourth argument
to NRmax is omitted then a numerical approximation to the Hessian is constructed.
What is ultimately required in Newton–Raphson is the negative inverse of the Hessian. Note that
if you give the optional fourth argument, your function should compute the negative Hessian, but
should not invert it; NRmax takes care of inversion, with special handling for the case where the
matrix is not negative definite, which can happen far from the maximum.
Chapter 34. Numerical methods 317
Script 34.3 extends the Rosenbrock example, using NRmax with a function Rosen_hess to compute
the Hessian. The functions Rosenbrock and Rosen_grad are just the same as in Listing 34.2 and
are omitted for brevity.
matrix theta = { 0, 0 }
matrix grad = { 0, 0 }
matrix H = zeros(2, 2)
set max_verbose 1
M = NRmax(&theta, Rosenbrock(theta), Rosen_grad(&grad, theta),
Rosen_hess(&H, theta))
print theta
print grad
The idea behind Newton–Raphson is to exploit a quadratic approximation to the maximand, under
the assumption that it is concave. If this is true, the method is very effective. However, if the
algorithm happens to evaluate the function at a point where the Hessian is not negative definite,
things may go wrong. Script 34.4 exemplifies this by using a normal density, which is concave in the
interval (−1, 1) and convex elsewhere. If the algorithm is started from within the interval everything
goes well and NR is (slightly) more effective than BFGS. If, however, the Hessian is positive at the
starting point BFGS converges with only little more difficulty, while Newton–Raphson fails.
set max_verbose 1
x = {0.75}
A = BFGSmax(&x, ND(x))
x = {0.75}
A = NRmax(&x, ND(x))
x = {1.5}
A = BFGSmax(&x, ND(x))
x = {1.5}
A = NRmax(&x, ND(x))
set max_verbose on
m = {5, 0, 10}
y = GSSmax(&m, g(m[1]))
printf "\nf(%g) = %g\n", m[1], y
The output is
1: bracket={3.81966,6.18034}, values={8.18747,1.59001}
2: bracket={2.36068,3.81966}, values={17.1118,8.18747}
3: bracket={1.45898,2.36068}, values={20.4841,17.1118}
4: bracket={0.901699,1.45898}, values={17.3764,20.4841}
...
20: bracket={1.50017,1.50042}, values={20.4958,20.4958}
21: bracket={1.50001,1.50017}, values={20.4958,20.4958}
f(1.49996) = 20.4958
As you can see from the output, the bracket shrinks progressively; the center of the interval when
the algorithm stops is x = 1.49996; figure 34.1 gives a pictorial representation of the process,
where the blue line is the function to maximize and the red segments are the successive choices
for the bracket.
Simulated Annealing
Simulated annealing—as implemented by the gretl function simann— is not a full-blown maxi-
mization method in its own right, but can be a useful auxiliary tool in problems where convergence
depends sensitively on the initial values of the parameters. The idea is that you supply initial values
and the simulated annealing mechanism tries to improve on them via controlled randomization.
The simann function takes up to three arguments. The first two (required) are the same as for
BFGSmax and NRmax: an initial parameter vector and a function that computes the maximand. The
Chapter 34. Numerical methods 319
25
20
15
10
0
0 1 2 3 4 5 6 7
optional third argument is a positive integer giving the maximum number of iterations, n, which
defaults to 1024.
Starting from the specified point in the parameter space, for each of n iterations we select at
random a new point within a certain radius of the previous one and determine the value of the
criterion at the new point. If the criterion is higher we jump to the new point; otherwise, we jump
with probability P (and remain at the previous point with probability 1 − P ). As the iterations
proceed, the system gradually “cools”— that is, the radius of the random perturbation is reduced,
as is the probability of making a jump when the criterion fails to increase.
In the course of this procedure n + 1 points in the parameter space are evaluated: call them θi , i =
0, . . . , n, where θ0 is the initial value given by the user. Let θ ∗ denote the “best” point among
θ1 , . . . , θn (highest criterion value). The value written into the parameter vector on completion is
then θ ∗ if θ ∗ is better than θ0 , otherwise θn . In other words, failing an actual improvement in
the criterion, simann randomizes the starting point, which may be helpful in tricky optimization
problems.
Listing 34.5 shows simann at work as a helper for BFGSmax in finding the maximum of a bimodal
function. Unaided, BFGSmax requires 60 function evaluations and 55 evaluations of the gradient,
while after simulated annealing the maximum is found with 7 function evaluations and 6 evalua-
tions of the gradient.1
Nelder–Mead
The Nelder–Mead derivative-free simplex maximizer (also known as the “amoeba” algorithm) is
implemented by the function NMmax. The argument list of this function is essentially the same as
for simann: the required arguments are an initial parameter vector and a function-call to compute
the maximand, while an optional third argument can be used to set the maximum number of
function evaluations (default value: 2000).
This method is unlikely to produce as close an approximation to the “true” optimum as derivative-
based methods such as BFGS and Newton–Raphson, but it is more robust than the latter. It may
succeed in some cases where derivative-based methods fail, and it may be useful, like simann, for
improving the starting point for an optimization problem so that a derivative-based method can
then take over successfully.
NMmax includes an internal “convergence” check — namely, verification that the best value achieved
for the objective function at termination of the algorithm is at least a local optimum — but by de-
1 Your mileage may vary: these figures are somewhat compiler- and machine-dependent.
Chapter 34. Numerical methods 320
scalar k = 2
matrix A = 0.1 * I(k)
matrix x0 = {3; -5}
x = x0
u = BFGSmax(&x, bimodal(x, A))
print x
x = x0
u = simann(&x, bimodal(x, A), 1000)
print x
u = BFGSmax(&x, bimodal(x, A))
print x
fault it doesn’t flag an error if this condition is not satisfied. This permits a mode of usage where
you set a fairly tight budget of function evaluations (for example, 200) and just take any improve-
ment in the objective function that is available, without worrying about whether an optimum has
truly been reached. However, if you want the convergence check to be enforced you can flag this by
setting a negative value for the maximum function evaluations argument; in that case the absolute
value of the argument is taken and an error is provoked on non-convergence.
If the task for this function is actually minimization, you can either have the function-call return
the negative of the actual criterion or, if you prefer, call NMmax under the alias NMmin.
Here is an example of use: minimization of the Powell quartic function, which is problematic for
BFGS. (The true minimum is zero, obtained for x a 4-vector of zeros.)
where we assume that SumOC is a user-defined function with the following structure:
This may come in handy in several cases: for example, if you use BFGSmax to estimate a model, you
may wish to calculate a numerical approximation to the relevant Jacobian to construct a covariance
matrix for your estimates.
Another example is the delta method: if you have a consistent estimator of a vector of parameters
θ̂, and a consistent estimate of its covariance matrix Σ, you may need to compute estimates for a
nonlinear continuous transformation ψ = g(θ). In this case, a standard result in asymptotic theory
is that
p
ψ̂ = g(θ̂) −p→ ψ = g(θ)
θ̂ −→ θ
=⇒
√T θ̂ − θ −d→ N(0, Σ) √T ψ̂ − ψ −d→ N(0, JΣJ 0 )
∂g(x)
where T is the sample size and J is the Jacobian ∂x .
x=θ
Script 34.6 exemplifies such a case: the example is taken from Greene (2003), section 9.3.1. The
slight differences between the results reported in the original source and what gretl returns are due
to the fact that the Jacobian is computed numerically, rather than analytically as in the book.
On the subject of numerical versus analytical derivatives, one may wonder what difference it makes
to use one method or another. Simply put, the answer is: analytical derivatives may be painful to
derive and to translate into code, but in most cases they are much faster than using fdjac; as a
consequence, if you need to use derivatives as part of an algorithm that requires iteration (such
as numerical optimization, or a Monte Carlo experiment), you’ll definitely want to use analytical
derivatives.
Analytical derivatives are also, in most cases, more precise than numerical ones, but this advantage
may or may not be negligible in practice depending on the practical details: the two fundamental
aspects to take in consideration are nonlinearity and machine precision.
As an example, consider the derivative of a highly nonlinear function such as the matrix inverse. In
order to keep the example simple, let’s focus on 2 × 2 matrices and define the function
which, given vec(A), returns vec(A−1 ). As is well known (see for example Magnus and Neudecker
(1988)),
∂vec(A−1 )
= −(A−1 )0 ⊗ (A−1 ),
∂vec(A)
Chapter 34. Numerical methods 322
matrix Y = realdpi[2000:4]
matrix theta = $coeff
matrix V = $vcv
Using the fdjac function to obtain the same result is even easier: you just invoke it like
fdjac(a, "vecinv(a)")
a = {2; 1; 1; 1}
ia = vecinv(a)
ag = grad(a)
ng = fdjac(a, "vecinv(a)")
dg = ag - ng
print ag ng dg
gives
ag (4 x 4)
-1 1 1 -1
1 -2 -1 2
1 -1 -2 2
-1 2 2 -4
ng (4 x 4)
-1 1 1 -1
1 -2 -1 2
1 -1 -2 2
-1 2 2 -4
dg (4 x 4)
ag (4 x 4)
ng (4 x 4)
dg (4 x 4)
ag (4 x 4)
ng (4 x 4)
dg (4 x 4)
Computing a Hessian
In principle, you can use the fdjac function repeatedly to evaluate higher-order derivatives. How-
ever, you should be aware that you may encounter rather serious numerical issues. For this reason,
gretl provides a function, similar to fdjac, for computing a Hessian directly, called numhess.
An example follows:
set fdjac_quality 2
H0 = fdjac(x, "coljac(x, A)")
H1 = numhess(x, "quad(x, A)")
printf "\n%14.8f", H0
printf "\n%14.8f", H1
In this example, we use the function quad for evaluating a simple quadratic form f (x) = 12 x 0 Ax,
for which the Hessian is trivially the matrix A. The matrix H0 is obtained by using fdjac twice
(at its best quality, that is with Richardson extrapolation on); the matrix H1, instead, is obtained
directly via the numhess function. The result of the above script is:
22.99788411 -10.99983724
-10.99902343 117.99161784
23.00000000 -11.00000000
-11.00000000 118.00000000
which makes it apparent how much better the second result is (note that H0 is not even symmetric,
as it should be).
where the q “quadrature points” x1 , x2 , . . . , xq are suitably chosen and the corresponding weights
w(xi ) ensure that the approximation is optimal.
The quadtable function provides quadrature points and weights for the following list of problems:
The quadtable function returns a q × 2 matrix with the xi points in the first column and the
corresponding weights wi in the second, so in practice you compute the desired integral of f (x)
by applying that function to the first column and then computing the inner product of the result
by the second column.
In econometrics, Gaussian quadrature is most commonly used for evaluating expectations of func-
tions of Gaussian random variables, that is, integrals of the form
Z∞
E [f (X)] = f (x)ϕ(x)dx
−∞
Chapter 34. Numerical methods 326
For example, consider the following script, in which we evaluate E(eX ) and E[cos(X)] via Gaussian
quadrature.
q = 10
Q = quadtable(q)
x = exp(sqrt(2) * Q[,1])
w = Q[,2] ./ sqrt($pi)
printf "E(exp(x)) = %16.14f (exact = %16.14f)\n", x’w, exp(0.5)
x = cos(sqrt(2) * Q[,1])
w = Q[,2] ./ sqrt($pi)
printf "E(cos(x)) = %16.14f (exact = %16.14f)\n", x’w, exp(-0.5)
The result is
C = E [f (X)] X ∼ N(µ, σ 2 )
Then you can use the syntax quadtable(q, 1, m, s), where the 1 indicates Gauss–Hermite inte-
gration and the two final arguments give the mean and standard deviation. So for example if you
want to compute E(X 2 ), where X ∼ N(1, 0.25), you could use
which yields
This chapter deals with models for dependent variables that are discrete or censored or otherwise
limited (as in event counts or durations, which must be positive) and that therefore call for estima-
tion methods other than the classical linear model. We discuss several estimators (mostly based on
the Maximum Likelihood principle), adding some details and examples to complement the material
on these methods in the Gretl Command Reference.
where zi is commonly known as the index function. Note that in this case the coefficients βj cannot
be interpreted as the partial derivatives of E(yi |xi ) with respect to xij . However, for a given value
of xi it is possible to compute the vector of “slopes”, that is
∂F (z)
slopej (x̄) =
∂xj z=z̄
Gretl automatically computes the slopes, setting each explanatory variable at its sample mean.
Another, equivalent way of thinking about this model is in terms of an unobserved variable yi∗
which can be described thus:
k
X
yi∗ = xij βj + εi = zi + εi (35.5)
j=1
327
Chapter 35. Discrete and censored dependent variables 328
Both the probit and logit model are estimated in gretl via maximum likelihood, where the log-
likelihood can be written as
X X
L(β) = ln[1 − F (zi )] + ln F (zi ), (35.6)
yi =0 yi =1
which is always negative, since 0 < F (·) < 1. Since the score equations do not have a closed form
solution, numerical optimization is used. However, in most cases this is totally transparent to the
user, since usually only a few iterations are needed to ensure convergence. The --verbose switch
can be used to track the maximization algorithm.
open greene19_1
As an example, we reproduce the results given in chapter 21 of Greene (2000), where the effective-
ness of a program for teaching economics is evaluated by the improvements of students’ grades.
Running the code in example 35.1 gives the output reported in Table 35.1. Note that for the probit
model a conditional moment test on skewness and kurtosis (Bera, Jarque and Lee, 1984) is printed
automatically as a test for normality.
In this context, the $uhat accessor function takes a special meaning: it returns generalized resid-
uals as defined in Gourieroux, Monfort, Renault and Trognon (1987), which can be interpreted as
unbiased estimators of the latent disturbances εi . These are defined as
yi − P̂i for the logit model
ui = (35.7)
yi · φ(ẑi ) − (1 − yi ) · φ(ẑi ) for the probit model
Φ(ẑi ) 1−Φ(ẑi )
Among other uses, generalized residuals are often used for diagnostic purposes. For example, it is
very easy to set up an omitted variables test equivalent to the familiar LM test in the context of a
linear regression; example 35.2 shows how to perform a variable addition test.
open greene19_1
Predicted
0 1
Actual 0 18 3
1 3 8
Predicted
0 1
Actual 0 18 3
1 3 8
why this problem arises is easy to see by considering equation (35.6): if for some vector β and scalar
k it’s the case that zi < k whenever yi = 0 and zi > k whenever yi = 1, the same thing is true
for any multiple of β. Hence, L(β) can be made arbitrarily close to 0 simply by choosing enormous
values for β. As a consequence, the log-likelihood has no maximum, despite being bounded.
Gretl has a mechanism for preventing the algorithm from iterating endlessly in search of a non-
existent maximum. One sub-case of interest is when the perfect prediction problem arises because
of a single binary explanatory variable. In this case, the offending variable is dropped from the
model and estimation proceeds with the reduced specification. Nevertheless, it may happen that
no single “perfect classifier” exists among the regressors, in which case estimation is simply impos-
sible and the algorithm stops with an error. This behavior is triggered during the iteration process
if
max zi < min zi
i:yi =0 i:yi =1
If this happens, unless your model is trivially mis-specified (like predicting if a country is an oil
exporter on the basis of oil revenues), it is normally a small-sample problem: you probably just
don’t have enough data to estimate your model. You may want to drop some of your explanatory
variables.
This problem is well analyzed in Stokes (2004); the results therein are replicated in the example
script murder_rates.inp.
y ∗ = Xβ + ε = z + ε
y =0 if y ∗ ≤ α1
y =1 if α1 < y ∗ ≤ α2
..
.
y =J if y ∗ > αJ
For example, if the response takes on three values there will be two such cut points, α1 and α2 .
The probability that individual i exhibits response j, conditional on the characteristics xi , is then
given by
∗
P (y ≤ α1 | xi ) = F (α1 − zi ) for j = 0
P (yi = j | xi ) = P (αj < y ∗ ≤ αj+1 | xi ) = F (αj+1 − zi ) − F (αj − zi ) for 0 < j < J (35.8)
P (y ∗ > α | x ) = 1 − F (α − z )
for j = J
J i J i
The unknown parameters αj are estimated jointly with the βs via maximum likelihood. The α̂j
estimates are reported by gretl as cut1, cut2 and so on. For the probit variant, a conditional
moment test for normality constructed in the spirit of Chesher and Irish (1987) is also included.
Note that the αj parameters can be shifted arbitrarily by adding a constant to zi , so the model is
under-identified if there is some linear combination of the explanatory variables which is constant.
The most obvious case in which this occurs is when the model contains a constant term; for this
Chapter 35. Discrete and censored dependent variables 331
reason, gretl drops automatically the intercept if present. However, it may happen that the user in-
adventently specifies a list of regressors that may be combined in such a way to produce a constant
(for example, by using a full set of dummy variables for a discrete factor). If this happens, gretl will
also drop any offending regressors.
In order to apply these models in gretl, the dependent variable must either take on only non-
negative integer values, or be explicitly marked as discrete. (In case the variable has non-integer
values, it will be recoded internally.) Note that gretl does not provide a separate command for
ordered models: the logit and probit commands automatically estimate the ordered version if
the dependent variable is acceptable, but not binary.
Listing 35.3 reproduces the results presented in section 15.10 of Wooldridge (2002a). The question
of interest in this analysis is what difference it makes, to the allocation of assets in pension funds,
whether individual plan participants have a choice in the matter. The response variable is an ordinal
measure of the weight of stocks in the pension portfolio. Having reported the results of estimation
of the ordered model, Wooldridge illustrates the effect of the choice variable by reference to an
“average” participant. The example script shows how one can compute this effect in gretl.
After estimating ordered models, the $uhat accessor yields generalized residuals as in binary mod-
els; additionally, the $yhat accessor function returns ẑi , so it is possible to compute an unbiased
estimator of the latent variable yi∗ simply by adding the two together.
exp(xi βk )
P (yi = k|xi ) = Pp
j=0 exp(xi βj )
For the purpose of identification one of the outcomes must be taken as the “baseline”; it is usually
assumed that β0 = 0, in which case
exp(xi βk )
P (yi = k|xi ) = Pp
1+ j=1 exp(xi βj )
and
1
P (yi = 0|xi ) = Pp .
1+ j=1 exp(xi βj )
Listing 35.4 reproduces Table 15.2 in Wooldridge (2002a), based on data on career choice from
Keane and Wolpin (1997). The dependent variable is the occupational status of an individual (0 = in
school; 1 = not in school and not working; 2 = working), and the explanatory variables are education
and work experience (linear and square) plus a “black” binary variable. The full data set is a panel;
here the analysis is confined to a cross-section for 1987.
k1
X
∗ ∗
y1,i = xij βj + ε1,i y1,i = 1 ⇐⇒ y1,i >0 (35.9)
j=1
Chapter 35. Discrete and censored dependent variables 332
/*
Replicate the results in Wooldridge, Econometric Analysis of Cross
Section and Panel Data, section 15.10, using pension-plan data from
Papke (AER, 1998).
open pension.gdt
k = $ncoeff
matrix b = $coeff[1:k-2]
a1 = $coeff[k-1]
a2 = $coeff[k]
/*
Wooldridge illustrates the ’choice’ effect in the ordered probit
by reference to a single, non-black male aged 60, with 13.5 years
of education, income in the range $50K - $75K and wealth of $200K,
participating in a plan with profit sharing.
*/
matrix X = {60, 13.5, 0, 0, 0, 0, 0, 0, 1, 0, 0, 200, 1}
# with ’choice’ = 0
scalar Xb = (0 ~ X) * b
P0 = cdf(N, a1 - Xb)
P50 = cdf(N, a2 - Xb) - P0
P100 = 1 - cdf(N, a2 - Xb)
E0 = 50 * P50 + 100 * P100
# with ’choice’ = 1
Xb = (1 ~ X) * b
P0 = cdf(N, a1 - Xb)
P50 = cdf(N, a2 - Xb) - P0
P100 = 1 - cdf(N, a2 - Xb)
E1 = 50 * P50 + 100 * P100
k2
X
∗ ∗
y2,i = zij γj + ε2,i y2,i = 1 ⇐⇒ y2,i >0 (35.10)
j=1
" # " !#
ε2,i 1 ρ
∼ N 0, (35.11)
ε2,i ρ 1
• The explanatory variables for the first equation x and for the second equation z may overlap
FIXME: expand.
as exemplified in the reprobit.inp sample script. The numerical technique used for this particular
estimator is Gauss-Hermite quadrature, which we’ll now briefly describe. Generalizing equation
(35.5) to a panel context, we get
k
X
∗
yi,t = xijt βj + αi + εi,t = zi,t + ωi,t (35.12)
j=1
in which we assume that the individual effect, αi , and the disturbance term, εi,t , are mutually
independent zero-mean Gaussian random variables. The composite error term, ωi,t = αi + εi,t , is
therefore a normal r. v. with mean zero and variance 1 + σα2 . Because of the individual effect, αi ,
observations for the same unit are not independent; the likelihood therefore has to be evaluated
on a per-unit basis, as
`i = log P yi,1 , yi,2 , . . . , yi,T .
and there’s no way to write the above as a product of individual terms.
However, the above probability could be written as a product if we were to treat αi as a constant;
in that case we would have
T
X x ijt β j + αi
`i |αi = Φ (2yi,t − 1) q
t=1 1 + σα2
The technique known as Gauss–Hermite quadrature is simply a way of approximating the above
integral via a sum of carefully chosen terms:2
m
X
`i ' (`i |αi = nk )wk
k=1
where the numbers nk and wk are known as quadrature points and weights, respectively. Of course,
accuracy improves with higher values of m, but so does CPU usage. Note that this technique can
also be used in more general cases by using the quadtable() function and the mle command via
the apparatus described in chapter 23. Here, however, the calculations were hard-coded in C for
maximal speed and efficiency.
Experience shows that a reasonable compromise can be achieved in most cases by choosing m in
the order of 20 or so; gretl uses 32 as a default value, but this can be changed via the --quadpoints
option, as in
where εi ∼ N(0, σ 2 ). If yi∗ were observable, the model’s parameters could be estimated via ordinary
least squares. On the contrary, suppose that we observe yi , defined as
a
for yi∗ ≤ a
yi = ∗
yi for a < yi∗ < b (35.13)
b ∗
for y ≥ b
i
In most cases found in the applied literature, a = 0 and b = ∞, so in practice negative values of yi∗
are not observed and are replaced by zeros.
In this case, regressing yi on the xi s does not yield consistent estimates of the parameters β,
Pk
because the conditional mean E(yi |xi ) is not equal to j=1 xij βj . It can be shown that restricting
the sample to non-zero observations would not yield consistent estimates either. The solution is to
estimate the parameters via maximum likelihood. The syntax is simply
As usual, progress of the maximization algorithm can be tracked via the --verbose switch, while
$uhat returns the generalized residuals. Note that in this case the generalized residual is defined
as ûi = E(εi |yi = 0) for censored observations, so the familiar equality ûi = yi − ŷi only holds for
uncensored observations, that is, when yi > 0.
An important difference between the Tobit estimator and OLS is that the consequences of non-
normality of the disturbance term are much more severe: non-normality implies inconsistency for
the Tobit estimator. For this reason, the output for the Tobit model includes the Chesher and Irish
(1987) normality test by default.
The general case in which a is nonzero and/or b is finite can be handled by using the options
--llimit and --rlimit. So, for example,
2 Some have suggested using a more refined method called adaptive Gauss-Hermite quadrature; this is not imple-
mented in gretl.
Chapter 35. Discrete and censored dependent variables 336
yi∗ = xi β + i
but we only know that mi ≤ yi∗ ≤ Mi , where the interval may be left- or right-unbounded (but
not both). If mi = Mi , we effectively observe yi∗ and no information loss occurs. In practice, each
observation belongs to one of four categories:
2. right-unbounded, when Mi = ∞,
It is interesting to note that this model bears similarities to other models in several special cases:
• When all observations are point observations the model trivially reduces to the ordinary linear
regression model.
• The interval model could be thought of as an ordered probit model (see 35.2) in which the cut
points (the αj coefficients in eq. 35.8) are observed and don’t need to be estimated.
• The Tobit model (see 35.6) is a special case of the interval model in which mi and Mi do not
depend on i, that is, the censoring limits are the same for all observations. As a matter of
fact, gretl’s tobit command is handled internally as a special case of the interval model.
The gretl command intreg estimates interval models by maximum likelihood, assuming normality
of the disturbance term i . Its syntax is
where minvar contains the mi series, with NAs for left-unbounded observations, and maxvar con-
tains Mi , with NAs for right-unbounded observations. By default, standard errors are computed
using the negative inverse of the Hessian. If the --robust flag is given, then QML or Huber–White
standard errors are calculated instead. In this case the estimated covariance matrix is a “sandwich”
of the inverse of the estimated Hessian and the outer product of the gradient.
If the model specification contains regressors other than just a constant, the output includes a
chi-square statistic for testing the joint null hypothesis that none of these regressors has any
effect on the outcome. This is a Wald statistic based on the estimated covariance matrix. If you
wish to construct a likelihood ratio test, this is easily done by estimating both the full model
and the null model (containing only the constant), saving the log-likelihood in both cases via the
$lnl accessor, and then referring twice the difference between the two log-likelihoods to the chi-
square distribution with k degrees of freedom, where k is the number of additional regressors (see
the pvalue command in the Gretl Command Reference). Also included is a conditional moment
normality test, similar to those provided for the probit, ordered probit and Tobit models (see
above). An example is contained in the sample script wtp.inp, provided with the gretl distribution.
Chapter 35. Discrete and censored dependent variables 337
# estimate ystar
gen_resid = $uhat
yhat = $yhat + gen_resid
corr ystar yhat
sigma = 0.223273
Left-unbounded observations: 0
Right-unbounded observations: 0
Bounded observations: 100
Point observations: 0
...
As with the probit and Tobit models, after a model has been estimated the $uhat accessor returns
the generalized residual, which is an estimate of i : more precisely, it equals yi − xi β̂ for point
observations and E(i |mi , Mi , xi ) otherwise. Note that it is possible to compute an unbiased pre-
dictor of yi∗ by summing this estimate to xi β̂. Script 35.5 shows an example. As a further similarity
with Tobit, the interval regression model may deliver inconsistent estimates if the disturbances are
non-normal; hence, the Chesher and Irish (1987) test for normality is included by default here too.
Chapter 35. Discrete and censored dependent variables 338
In this context, the ♦ symbol indicates that for some observations we simply do not have data on
y: yi may be 0, or missing, or anything else. A dummy variable di is normally used to set censored
observations apart.
One of the most popular applications of this model in econometrics is a wage equation coupled
with a labor force participation equation: we only observe the wage for the employed. If yi∗ and si∗
were (conditionally) independent, there would be no reason not to use OLS for estimating equation
(35.14); otherwise, OLS does not yield consistent estimates of the parameters βj .
Since conditional independence between yi∗ and si∗ is equivalent to conditional independence be-
tween εi and ηi , one may model the co-dependence between εi and ηi as
εi = ληi + vi ;
substituting the above expression in (35.14), you obtain the model that is actually estimated:
k
X
yi = xij βj + λη̂i + vi ,
j=1
so the hypothesis that censoring does not matter is equivalent to the hypothesis H0 : λ = 0, which
can be easily tested.
The parameters can be estimated via maximum likelihood under the assumption of joint normality
of εi and ηi ; however, a widely used alternative method yields the so-called Heckit estimator, named
after Heckman (1979). The procedure can be briefly outlined as follows: first, a probit model is fit
on equation (35.15); next, the generalized residuals are inserted in equation (35.14) to correct for
the effect of sample selection.
Gretl provides the heckit command to carry out estimation; its syntax is
heckit y X ; d Z
where y is the dependent variable, X is a list of regressors, d is a dummy variable holding 1 for
uncensored observations and Z is a list of explanatory variables for the censoring equation.
Since in most cases maximum likelihood is the method of choice, by default gretl computes ML
estimates. The 2-step Heckit estimates can be obtained by using the --two-step option. After
estimation, the $uhat accessor contains the generalized residuals. As in the ordinary Tobit model,
the residuals equal the difference between actual and fitted yi only for uncensored observations
(those for which di = 1).
Listing 35.6 shows two estimates from the dataset used in Mroz (1987): the first one replicates
Table 22.7 in Greene (2003),3 while the second one replicates table 17.1 in Wooldridge (2002a).
3 Note that the estimates given by gretl do not coincide with those found in the printed volume. They do, however,
match those found on the errata web page for Greene’s book: http://pages.stern.nyu.edu/~wgreene/Text/Errata/
ERRATA5.htm.
Chapter 35. Discrete and censored dependent variables 339
open mroz87.gdt
# Greene’s specification
# Wooldridge’s specification
In some cases, an “offset” variable is needed. The number of occurrences of yi in a given time is
assumed to be strictly proportional to the offset variable ni . In the epidemiology literature, the
offset is known as “population at risk”. In this case, the model becomes
X
µi = ni exp xij βj
j
Another way to look at the offset variable is to consider its natural log as just another explanatory
variable whose coefficient is constrained to be one.
Estimation is carried out by maximum likelihood and follows the syntax
If an offset variable is needed, it has to be specified at the end of the command, separated from the
list of explanatory variables by a semicolon, as in
It should be noted that the poisson command does not use, internally, the same optimization
engines as most other gretl command, such as arma or tobit. As a consequence, some details may
differ: the --verbose option will yield different output and settings such as bfgs_toler will not
work.
Overdispersion
In the Poisson model, E(yi |xi ) = V (yi |xi ) = µi , that is, the conditional mean equals the conditional
variance by construction. In many cases, this feature is at odds with the data, as the conditional
variance is often larger than the mean; this phenomenon is called “overdispersion”. The output
from the poisson command includes a conditional moment test for overdispersion (as per David-
son and MacKinnon (2004), section 11.5), which is printed automatically after estimation.
Overdispersion can be attributed to unmodeled heterogeneity between individuals. Two data points
with the same observable characteristics xi = xj may differ because of some unobserved scale
factor si 6= sj so that
E(yi |xi , si ) = µi si 6= µj sj = E(yj |xj , sj )
even though µi = µj . In other words, yi is a Poisson random variable conditional on both xi and si ,
but since si is unobservable, the only thing we can we can use, P (yi |xi ), will not follow the Poisson
distribution.
It is often assumed that si can be represented as a gamma random variable with mean 1 and
variance α: the parameter α is estimated together with the vector β, and measures the degree of
heterogeneity between individuals.
In this case, the conditional probability for yi given xi can be shown to be
" #yi " #α−1
Γ (yi + α−1 ) µi α−1
P (yi |xi ) = (35.17)
Γ (α−1 )Γ (yi + 1) µi + α−1 µi + α−1
which is known as the “Negative Binomial Model”. The conditional mean is still E(yi |xi ) = µi , but
the variance equals V (yi |xi ) = µi (1 + µi α). The gretl command for this model is negbin depvar
indep.
• There is also a less used variant of the negative binomial model, in which the conditional vari-
ance is a scalar multiple of the conditional mean, that is V (yi |xi ) = µi (1 + γ). To distinguish
between the two, the model (35.17) is termed “Type 2”. Gretl implements model 1 via the
option --model1.
• A script which exemplifies the above models is included among gretl’s sample scripts, under
the name camtriv.inp.
FIXME: expand.
• From engineering, the “time to failure” of electronic or mechanical components: how long do,
say, computer hard drives last until they malfunction?
• From the medical realm: how does a new treatment affect the time from diagnosis of a certain
condition to exit from that condition (where “exit” might mean death or full recovery)?
In each case we may be interested in how the durations are distributed, and how they are affected
by relevant covariates. There are several approaches to this problem; the one we discuss here —
which is currently the only one supported by gretl — is estimation of a parametric model by means
of Maximum Likelihood. In this approach we hypothesize that the durations follow some definite
probability law and we seek to estimate the parameters of that law, factoring in the influence of
covariates.
We may express the density of the durations as f (t, X, θ), where t is the length of time in the state
in question, X is a matrix of covariates, and θ is a vector of parameters. The likelihood for a sample
of n observations indexed by i is then
n
Y
L= f (ti , xi , θ)
i=1
Rather than working with the density directly, however, it is standard practice to factor f (·) into
two components, namely a hazard function, λ, and a survivor function, S. The survivor function
gives the probability that a state lasts at least as long as t; it is therefore 1 − F (t, X, θ) where F
is the CDF corresponding to the density f (·). The hazard function addresses this question: given
that a state has persisted as long as t, what is the likelihood that it ends within a short increment
of time beyond t —that is, it ends between t and t + ∆? Taking the limit as ∆ goes to zero, we end
up with the ratio of the density to the survivor function:4
f (t, X, θ)
λ(t, X, θ) = (35.18)
S(t, X, θ)
so the log-likelihood can be written as
n
X n
X
`= log f (ti , xi , θ) = log λ(ti , xi , θ) + log S(ti , xi , θ) (35.19)
i=1 i=1
One point of interest is the shape of the hazard function, in particular its dependence (or not) on
time since the state began. If λ does not depend on t we say the process in question exhibits du-
ration independence: the probability of exiting the state at any given moment neither increases nor
decreases based simply on how long the state has persisted to date. The alternatives are positive
duration dependence (the likelihood of exiting the state rises, the longer the state has persisted)
or negative duration dependence (exit becomes less likely, the longer it has persisted). Finally, the
behavior of the hazard with respect to time need not be monotonic; some parameterizations allow
for this possibility and some do not.
Since durations are inherently positive the probability distribution used in modeling must respect
this requirement, giving a density of zero for t ≤ 0. Four common candidates are the exponential,
Weibull, log-logistic and log-normal, the Weibull being the most common choice. The table below
displays the density and the hazard function for each of these distributions as they are commonly
parameterized, written as functions of t alone. (φ and Φ denote, respectively, the Gaussian PDF
and CDF.)
The hazard is constant for the exponential distribution. For the Weibull, it is monotone increasing
in t if α > 1, or monotone decreasing for α < 1. (If α = 1 the Weibull collapses to the exponential.)
The log-logistic and log-normal distributions allow the hazard to vary with t in a non-monotonic
fashion.
Covariates are brought into the picture by allowing them to govern one of the parameters of the
density, so that durations are not identically distributed across cases. For example, when using
the log-normal distribution it is natural to make µ, the expected value of log t, depend on the
covariates, X. This is typically done via a linear index function: µ = Xβ.
Note that the expressions for the log-normal density and hazard contain the term (log t − µ)/σ .
Replacing µ with Xβ this becomes (log t − Xβ)/σ . As in Kalbfleisch and Prentice (2002), we define
a shorthand label for this term:
wi ≡ (log ti − xi β)/σ (35.20)
It turns out that this constitutes a useful simplifying change of variables for all of the distributions
discussed here. The interpretation of the scale factor, σ , in the expression above depends on the
distribution. For the log-normal, σ represents the standard deviation of log t; for the Weibull and
the log-logistic it corresponds to 1/α; and for the exponential it is fixed at unity. For distributions
other than the log-normal, Xβ corresponds to − log γ, or in other words γ = exp(−Xβ).
With this change of variables, the density and survivor functions may be written compactly as
follows (the exponential is the same as the Weibull).
In light of the above we may think of the generic parameter vector θ, as in f (t, X, θ), as composed of
the coefficients on the covariates, β, plus (in all cases but the exponential) the additional parameter
σ.
A complication in estimation of θ is posed by “incomplete spells”. That is, in some cases the state
in question may not have ended at the time the observation is made (e.g. some workers remain
unemployed, some components have not yet failed). If we use ti to denote the time from entering
the state to either (a) exiting the state or (b) the observation window closing, whichever comes first,
then all we know of the “right-censored” cases (b) is that the duration was at least as long as ti .
This can be handled by rewriting the the log-likelihood (compare 35.19) as
n
X
`i = δi log S (wi ) + (1 − δi ) − log σ + log f (wi ) (35.21)
i=1
where δi equals 1 for censored cases (incomplete spells), and 0 for complete observations. The
rationale for this is that the log-density equals the sum of the log hazard and the log survivor
function, but for the incomplete spells only the survivor function contributes to the likelihood. So
in (35.21) we are adding up the log survivor function alone for the incomplete cases, plus the full
log density for the completed cases.
where durat measures durations, 0 represents the constant (which is required for such models), X
is a named list of regressors, and cens is the censoring dummy.
By default the Weibull distribution is used; you can substitute any of the other three distribu-
tions discussed here by appending one of the option flags --exponential, --loglogistic or
--lognormal.
Interpreting the coefficients in a duration model requires some care, and we will work through
an illustrative case. The example comes from section 20.3 of Wooldridge (2002a) and it concerns
criminal recidivism.5 The data (filename recid.gdt) pertain to a sample of 1,445 convicts released
from prison between July 1, 1977 and June 30, 1978. The dependent variable is the time in months
until they are again arrested. The information was gathered retrospectively by examining records
in April 1984; the maximum possible length of observation is 81 months. Right-censoring is impor-
tant: when the date were compiled about 62 percent had not been rearrested. The dataset contains
several covariates, which are described in the data file; we will focus below on interpretation of the
married variable, a dummy which equals 1 if the respondent was married when imprisoned.
Listing 35.7 shows the gretl commands for Weibull and log-normal models along with most of the
output. Consider first the Weibull scale factor, σ . The estimate is 1.241 with a standard error of
0.048. (We don’t print a z score and p-value for this term since H0 : σ = 0 is not of interest.)
Recall that σ corresponds to 1/α; we can be confident that α is less than 1, so recidivism displays
negative duration dependence. This makes sense: it is plausible that if a past offender manages
to stay out of trouble for an extended period his risk of engaging in crime again diminishes. (The
exponential model would therefore not be appropriate in this case.)
On a priori grounds, however, we may doubt the monotonic decline in hazard that is implied by
the Weibull specification. Even if a person is liable to return to crime, it seems relatively unlikely
that he would do so straight out of prison. In the data, we find that only 2.6 percent of those
followed were rearrested within 3 months. The log-normal specification, which allows the hazard
to rise and then fall, may be more appropriate. Using the duration command again with the same
covariates but the --lognormal flag, we get a log-likelihood of −1597 as against −1633 for the
Weibull, confirming that the log-normal gives a better fit.
Let us now focus on the married coefficient, which is positive in both specifications but larger and
more sharply estimated in the log-normal variant. The first thing is to get the interpretation of the
sign right. Recall that Xβ enters negatively into the intermediate variable w (equation 35.20). The
Weibull hazard is λ(wi ) = ewi , so being married reduces the hazard of re-offending, or in other
words lengthens the expected duration out of prison. The same qualitative interpretation applies
for the log-normal.
To get a better sense of the married effect, it is useful to show its impact on the hazard across time.
We can do this by plotting the hazard for two values of the index function Xβ: in each case the
values of all the covariates other than married are set to their means (or some chosen values) while
married is set first to 0 then to 1. Listing 35.8 provides a script that does this, and the resulting
plots are shown in Figure 35.1. Note that when computing the hazards we need to multiply by the
Jacobian of the transformation from ti to wi = log(ti − xi β)/σ , namely 1/t. Note also that the
estimate of σ is available via the accessor $sigma, but it is also present as the last element in the
coefficient vector obtained via $coeff.
A further difference between the Weibull and log-normal specifications is illustrated in the plots.
The Weibull is an instance of a proportional hazard model. This means that for any sets of values of
the covariates, xi and xj , the ratio of the associated hazards is invariant with respect to duration. In
this example the Weibull hazard for unmarried individuals is always 1.1637 times that for married.
In the log-normal variant, on the other hand, this ratio gradually declines from 1.6703 at one month
to 1.1766 at 100 months.
5 Germán Rodríguez of Princeton University has a page discussing this example and displaying estimates from Stata
at http://data.princeton.edu/pop509/recid1.html.
Chapter 35. Discrete and censored dependent variables 344
Partial output:
Model 1: Duration (Weibull), using observations 1-1445
Dependent variable: durat
open recid.gdt -q
# Weibull variant
duration durat 0 X married ; cens
# coefficients on all Xs apart from married
matrix beta_w = $coeff[1:$ncoeff-2]
# married coefficient
scalar mc_w = $coeff[$ncoeff-1]
scalar s_w = $sigma
# Log-normal variant
duration durat 0 X married ; cens --lognormal
matrix beta_n = $coeff[1:$ncoeff-2]
scalar mc_n = $coeff[$ncoeff-1]
scalar s_n = $sigma
list allX = 0 X
# evaluate X\beta at means of all variables except marriage
scalar Xb_w = meanc({allX}) * beta_w
scalar Xb_n = meanc({allX}) * beta_n
loop t=1..100 -q
# first column, duration
mat_w[t, 1] = t
mat_n[t, 1] = t
wi_w = (log(t) - Xb_w)/s_w
wi_n = (log(t) - Xb_n)/s_n
# second col: hazard with married = 0
mat_w[t, 2] = (1/t) * exp(wi_w)
mat_n[t, 2] = (1/t) * pdf(z, wi_n) / cdf(z, -wi_n)
wi_w = (log(t) - (Xb_w + mc_w))/s_w
wi_n = (log(t) - (Xb_n + mc_n))/s_n
# third col: hazard with married = 1
mat_w[t, 3] = (1/t) * exp(wi_w)
mat_n[t, 3] = (1/t) * pdf(z, wi_n) / cdf(z, -wi_n)
endloop
Weibull
0.020
unmarried
0.018 married
0.016
0.014
0.012
0.010
0.008
0.006
0 20 40 60 80 100
months
Log-normal
0.020
unmarried
0.018 married
0.016
0.014
0.012
0.010
0.008
0.006
0 20 40 60 80 100
months
Figure 35.1: Recidivism hazard estimates for married and unmarried ex-convicts
Chapter 35. Discrete and censored dependent variables 347
The expression given for the log-logistic mean, however, is valid only for σ < 1; otherwise the
expectation is undefined, a point that is not noted in all software.6
Alternatively, if the --medians option is given, gretl’s duration command will produce conditional
medians as the content of $yhat. For the Weibull the median is exp(Xβ)(log 2)σ ; for the log-logistic
and log-normal it is just exp(Xβ).
The values we give for the accessor $uhat are generalized (Cox–Snell) residuals, computed as the
integrated hazard function, which equals the negative log of the survivor function:
Under the null of correct specification of the model these generalized residuals should follow the
unit exponential distribution, which has mean and variance both equal to 1 and density exp(−).
See chapter 18 of Cameron and Trivedi (2005) for further discussion.
6 The predict adjunct to the streg command in Stata 10, for example, gaily produces large negative values for the
Quantile regression
36.1 Introduction
In Ordinary Least Squares (OLS) regression, the fitted values, ŷi = Xi β̂, represent the conditional
mean of the dependent variable—conditional, that is, on the regression function and the values
of the independent variables. In median regression, by contrast and as the name implies, fitted
values represent the conditional median of the dependent variable. It turns out that the principle of
estimation for median regression is easily stated (though not so easily computed), namely, choose
β̂ so as to minimize the sum of absolute residuals. Hence the method is known as Least Absolute
Deviations or LAD. While the OLS problem has a straightforward analytical solution, LAD is a linear
programming problem.
Quantile regression is a generalization of median regression: the regression function predicts the
conditional τ-quantile of the dependent variable — for example the first quartile (τ = .25) or the
ninth decile (τ = .90).
If the classical conditions for the validity of OLS are satisfied — that is, if the error term is indepen-
dently and identically distributed, conditional on X — then quantile regression is redundant: all the
conditional quantiles of the dependent variable will march in lockstep with the conditional mean.
Conversely, if quantile regression reveals that the conditional quantiles behave in a manner quite
distinct from the conditional mean, this suggests that OLS estimation is problematic.
As of version 1.7.5, gretl offers quantile regression functionality (in addition to basic LAD regres-
sion, which has been available since early in gretl’s history via the lad command).1
where
• reglist is a standard gretl regression list (dependent variable followed by regressors, including
the constant if an intercept is wanted); and
• tau is the desired conditional quantile, in the range 0.01 to 0.99, given either as a numerical
value or the name of a pre-defined scalar variable (but see below for a further option).
Estimation is via the Frisch–Newton interior point solver (Portnoy and Koenker, 1997), which is sub-
stantially faster than the “traditional” Barrodale–Roberts (1974) simplex approach for large prob-
lems.
1 We gratefully acknowledge our borrowing from the quantreg package for GNU R (version 4.17). The core of the
quantreg package is composed of Fortran code written by Roger Koenker; this is accompanied by various driver and
auxiliary functions written in the R language by Koenker and Martin Mächler. The latter functions have been re-worked
in C for gretl. We have added some guards against potential numerical problems in small samples.
348
Chapter 36. Quantile regression 349
By default, standard errors are computed according to the asymptotic formula given by Koenker
and Bassett (1978). Alternatively, if the --robust option is given, we use the sandwich estimator
developed in Koenker and Zhao (1994).2
When the confidence intervals option is selected, the parameter estimates are calculated using
the Barrodale–Roberts method. This is simply because the Frisch–Newton code does not currently
support the calculation of confidence intervals.
Two further details. First, the mechanisms for generating confidence intervals for quantile esti-
mates require that the model has at least two regressors (including the constant). If the --intervals
option is given for a model containing only one regressor, an error is flagged. Second, when a model
is estimated in this mode, you can retrieve the confidence intervals using the accessor $coeff_ci.
This produces a k × 2 matrix, where k is the number of regressors. The lower bounds are in the
first column, the upper bounds in the second. See also section 36.5 below.
2 These correspond to the iid and nid options in R’s quantreg package, respectively.
Chapter 36. Quantile regression 350
Coefficient on income
0.75
Quantile estimates with 90% band
OLS estimate with 90% band
0.7
0.65
0.6
0.55
0.5
0.45
0.4
0.35
0.3
0 0.2 0.4 0.6 0.8 1
tau
The gretl GUI has an entry for Quantile Regression (under /Model/Robust estimation), and you can
select multiple quantiles there too. In that context, just give space-separated numerical values (as
per the predefined options, shown in a drop-down list).
When you estimate a model in this way most of the standard menu items in the model window
are disabled, but one extra item is available — graphs showing the τ sequence for a given coeffi-
cient in comparison with the OLS coefficient. An example is shown in Figure 36.1. This sort of
graph provides a simple means of judging whether quantile regression is redundant (OLS is fine) or
informative.
In the example shown—based on data on household income and food expenditure gathered by
Ernst Engel (1821–1896)—it seems clear that simple OLS regression is potentially misleading. The
“crossing” of the OLS estimate by the quantile estimates is very marked.
However, it is not always clear what implications should be drawn from this sort of conflict. With
the Engel data there are two issues to consider. First, Engel’s famous “law” claims an income-
elasticity of food consumption that is less than one, and talk of elasticities suggests a logarithmic
formulation of the model. Second, there are two apparently anomalous observations in the data
set: household 105 has the third-highest income but unexpectedly low expenditure on food (as
judged from a simple scatter plot), while household 138 (which also has unexpectedly low food
consumption) has much the highest income, almost twice that of the next highest.
With n = 235 it seems reasonable to consider dropping these observations. If we do so, and adopt
a log–log formulation, we get the plot shown in Figure 36.2. The quantile estimates still cross the
OLS estimate, but the “evidence against OLS” is much less compelling: the 90 percent confidence
bands of the respective estimates overlap at all the quantiles considered.
Coefficient on log(income)
0.96
0.94
0.92
0.9
0.88
0.86
0.84
0.82
0.8
0.78
Quantile estimates with 90% band
OLS estimate with 90% band
0.76
0 0.2 0.4 0.6 0.8 1
tau
Figure 36.2: Log–log regression; 2 observations dropped from full Engel data set.
This is OK for moderately large datasets (up to, say, a few thousand observations) but on very large
problems the simplex algorithm may become seriously bogged down. For example, Koenker and
Hallock (2001) present an analysis of the determinants of birth weights, using 198377 observations
and with 15 regressors. Generating confidence intervals via Barrodale–Roberts for a single value of
τ took about half an hour on a Lenovo Thinkpad T60p with 1.83GHz Intel Core 2 processor.
If you want confidence intervals in such cases, you are advised not to use the --intervals option,
but to compute them using the method of “plus or minus so many standard errors”. (One Frisch–
Newton run took about 8 seconds on the same machine, showing the superiority of the interior
point method.) The script below illustrates:
The matrix ci will contain the lower and upper bounds of the (symmetrical) 90 percent confidence
intervals.
To avoid a situation where gretl becomes unresponsive for a very long time we have set the maxi-
mum number of iterations for the Borrodale–Roberts algorithm to the (somewhat arbitrary) value
of 1000. We will experiment further with this, but for the meantime if you really want to use this
method on a large dataset, and don’t mind waiting for the results, you can increase the limit using
the set command with parameter rq_maxiter, as in
Nonparametric methods
The main focus of gretl is on parametric estimation, but we offer a selection of nonparametric
methods. The most basic of these
• various tests for difference in distribution (Sign test, Wilcoxon rank-sum test, Wilcoxon signed-
rank test);
Details on the above can be found by consulting the help for the commands difftest, runs, corr
and spearman. In the GUI program these items are found under the Tools menu and the Robust
estimation item under the Model menu.
In this chapter we concentrate on two relatively complex methods for nonparametric curve-fitting
and prediction, namely William Cleveland’s “loess” (also known as “lowess”) and the Nadaraya–
Watson estimator.
wk (xi ) = W (h−1
i (xk − xi ))
where hi is the distance between xi and its r th nearest neighbor, and W (·) is the tricube function,
(1 − |x|3 )3
(
for |x| < 1
W (x) =
0 for |x| ≥ 1
352
Chapter 37. Nonparametric methods 353
The local regression can be made robust via an adjustment based on the residuals, ei = yi − ŷi .
Robustness weights, δk , are defined by
δk = B(ek /6s)
where s is the median of the |ei | and B(·) is the bisquare function,
(1 − x 2 )2
(
for |x| < 1
B(x) =
0 for |x| ≥ 1
An illustration of loess is provided in Listing 37.1: we generate a series that has a deterministic
sine wave component overlaid with noise uniformly distributed on (−1, 1). Loess is then used to
retrieve a good approximation to the sine function. The resulting graph is shown in Figure 37.1.
nulldata 120
series x = index
scalar n = $nobs
series y = sin(2*pi*x/n) + uniform(-1, 1)
series yh = loess(y, x, 2, 0.75, 0)
gnuplot y yh x --output=display --with-lines=yh
2
loess fit
1.5
0.5
-0.5
-1
-1.5
-2
0 20 40 60 80 100 120
x
where Kh (·) is the so-called kernel function, which is usually some simple transform of a density
function that depends on a scalar, h, known as the bandwidth. The one used by gretl is
!
x2
Kh (x) = exp −
2h
for |x| < τ and zero otherwise. Larger values of h produce a smoother function. The scalar τ,
known as the trim parameter, is used to prevent numerical problems when the kernel function is
evaluated too far away from zero.
A common variant of Nadaraya–Watson is the so-called “leave-one-out” estimator, which omits the
i-th observation when evaluating m(xi ). The formula therefore becomes
P
j6=i yj · Kh (xi − xj )
m(xi ) = P
j6=i Kh (xi − xj )
This makes the estimator more robust numerically and its usage is often advised for inference
purposes.
The nadarwat() function in gretl takes up to five arguments as follows: the dependent series y,
the independent series x, the bandwidth h, a Boolean switch to turn on “leave-one-out”, and a value
for the trim parameter τ, expressed as a multiple of h. The last three arguments are optional; if
they are omitted the default values are, respectively, an automatic data-determined value for h (see
below), leave-one-out not activated, and τ = 4. The default value of τ offers a relatively safe guard
against numerical problems; in some cases a larger τ may produce more sensible values in regions
of X with sparse support.
Choice of bandwidth
As mentioned above, larger values of h lead to a smoother m(·) function; smaller values make
the m(·) function follow the yi values more closely, so that the function appears more “jagged”.
In fact, as h → ∞, m(xi ) → Ȳ ; on the contrary, if h → 0, observations for which xi 6= X are not
taken into account at all when computing m(X). Also, the statistical properties of m(·) vary with
h: its variance can be shown to be decreasing in h, while its squared bias is increasing in h. It
can be shown that choosing h ∼ n−1/5 minimizes the RMSE, so that value is customarily taken as a
reference point.
If the argument h is omitted or set to 0, gretl uses the following data-determined value:
r
h = 0.9 · min s, · n−1/5
1.349
60
m0
m1
m2
55
50
HA
45
40
35
30
30 35 40 45 50 55 60
WA
Figure 37.2: Nadaraya–Watson example for several choices of the bandwidth parameter
Chapter 37. Nonparametric methods 356
If you need a point estimate of m(X) for some value of X which is not present among the valid
observations of your dependent variable, you may want to add some “fake” observations to your
dataset in which y is missing and x contains the values you want m(x) evaluated at. For example,
the following script evaluates m(x) at regular intervals between −2.0 and 2.0:
nulldata 120
set seed 120496
x m
Technical details
357
Chapter 38
38.1 Introduction
TEX — initially developed by Donald Knuth of Stanford University and since enhanced by hundreds
of contributors around the world — is the gold standard of scientific typesetting. Gretl provides
various hooks that enable you to preview and print econometric results using the TEX engine, and
to save output in a form suitable for further processing with TEX.
This chapter explains the finer points of gretl’s TEX-related functionality. The next section describes
the relevant menu items; section 38.3 discusses ways of fine-tuning TEX output; and section 38.4
gives some pointers on installing (and learning) TEX if you do not already have it on your computer.
(Just to be clear: TEX is not included with the gretl distribution; it is a separate package, including
several programs and a large number of supporting files.)
Before proceeding, however, it may be useful to set out briefly the stages of production of a final
document using TEX. For the most part you don’t have to worry about these details, since, in regard
to previewing at any rate, gretl handles them for you. But having some grasp of what is going on
behind the scences will enable you to understand your options better.
The first step is the creation of a plain text “source” file, containing the text or mathematics to
be typset, interspersed with mark-up that defines how it should be formatted. The second step
is to run the source through a processing engine that does the actual formatting. Typically this a
program called pdflatex that generates PDF output.1 (In times gone by it was a program called latex
that generated so-called DVI (device-independent) output.)
So gretl calls pdflatex to process the source file. On MS Windows and Mac OS X, gretl expects the
operating system to find the default viewer for PDF output. On GNU/Linux you can specify your
preferred PDF viewer via the menu item “Tools, Preferences, General,” under the “Programs” tab.
ENROLL
Æ = 0.241105 + 0.223530 CATHOL − 0.00338200 PUPIL − 0.152643 WHITE
(0.066022) (0.04597) (0.0027196) (0.040706)
358
Chapter 38. Gretl and TEX 359
The distinction between the “Copy” and “Save” options (for both tabular and equation) is twofold.
First, “Copy” puts the TEX source on the clipboard while with “Save” you are prompted for the name
of a file into which the source should be saved. Second, with “Copy” the material is copied as a
“fragment” while with “Save” it is written as a complete file. The point is that a well-formed TEX
source file must have a header that defines the documentclass (article, report, book or whatever)
and tags that say \begin{document} and \end{document}. This material is included when you do
“Save” but not when you do “Copy”, since in the latter case the expectation is that you will paste
the data into an existing TEX source file that already has the relevant apparatus in place.
The items under “Equation options” should be self-explanatory: when printing the model in equa-
tion form, do you want standard errors or t-ratios displayed in parentheses under the parameter
estimates? The default is to show standard errors; if you want t-ratios, select that item.
Other windows
Several other sorts of output windows also have TEX preview, copy and save enabled. In the case of
windows having a graphical toolbar, look for the TEX button. Figure 38.2 shows this icon (second
from the right on the toolbar) along with the dialog that appears when you press the button.
One aspect of gretl’s TEX support that is likely to be particularly useful for publication purposes is
the ability to produce a typeset version of the “model table” (see section 3.4). An example of this is
shown in Table 38.2.
\documentclass[11pt]{article}
\usepackage[utf8]{inputenc}
Chapter 38. Gretl and TEX 361
OLS estimates
Dependent variable: ENROLL
ADMEXP −0.1551
(0.1342)
n 51 51 51
2
R̄ 0.4502 0.4462 0.2956
` 96.09 95.36 88.69
\usepackage{amsmath}
\usepackage{dcolumn,longtable}
\begin{document}
\thispagestyle{empty}
Note that the amsmath and dcolumn packages are required. (For some sorts of output the longtable
package is also needed.) Beyond that you can, for instance, change the type size or the font by al-
tering the documentclass declaration or including an alternative font package.
In addition, if you wish to typeset gretl output in more than one language, you can set up per-
language preamble files. A “localized” preamble file is identified by a name of the form gretlpre_xx.tex,
where xx is replaced by the first two letters of the current setting of the LANG environment vari-
able. For example, if you are running the program in Polish, using LANG=pl_PL, then gretl will do
the following when writing the preamble for a TEX source file.
1. Look for a file named gretlpre_pl.tex in the gretl user directory. If this is not found, then
2. look for a file named gretlpre.tex in the gretl user directory. If this is not found, then
3. use the default preamble.
Conversely, suppose you usually run gretl in a language other than English, and have a suitable
gretlpre.tex file in place for your native language. If on some occasions you want to produce TEX
output in English, then you could create an additional file gretlpre_en.tex: this file will be used
for the preamble when gretl is run with a language setting of, say, en_US.
Command-line options
After estimating a model via a script— or interactively via the gretl console or using the command-
line program gretlcli—you can use the commands tabprint or eqnprint to print the model to
file in tabular format or equation format respectively. These options are explained in the Gretl
Command Reference.
If you wish alter the appearance of gretl’s tabular output for models in the context of the tabprint
command, you can specify a custom row format using the --format flag. The format string must
be enclosed in double quotes and must be tied to the flag with an equals sign. The pattern for the
format string is as follows. There are four fields, representing the coefficient, standard error, t-
ratio and p-value respectively. These fields should be separated by vertical bars; they may contain
a printf-type specification for the formatting of the numeric value in question, or may be left
blank to suppress the printing of that column (subject to the constraint that you can’t leave all the
columns blank). Here are a few examples:
--format="%.4f|%.4f|%.4f|%.4f"
--format="%.4f|%.4f|%.3f|"
--format="%.5f|%.4f||%.4f"
--format="%.8g|%.8g||%.4f"
The first of these specifications prints the values in all columns using 4 decimal places. The second
suppresses the p-value and prints the t-ratio to 3 places. The third omits the t-ratio. The last one
again omits the t, and prints both coefficient and standard error to 8 significant figures.
Once you set a custom format in this way, it is remembered and used for the duration of the gretl
session. To revert to the default formatting you can use the special variant --format=default.
Further editing
Once you have pasted gretl’s TEX output into your own document, or saved it to file and opened it
in an editor, you can of course modify the material in any wish you wish. In some cases, machine-
generated TEX is hard to understand, but gretl’s output is intended to be human-readable and
Chapter 38. Gretl and TEX 363
-editable. In addition, it does not use any non-standard style packages. Besides the standard LATEX
document classes, the only files needed are, as noted above, the amsmath, dcolumn and longtable
packages. These should be included in any reasonably full TEX implementation.
Gretl and R
39.1 Introduction
R is, by far, the largest free statistical project.1 Like gretl, it is a GNU project and the two have a
lot in common; however, gretl’s approach focuses on ease of use much more than R, which instead
aims to encompass the widest possible range of statistical procedures.
As is natural in the free software ecosystem, we don’t view ourselves as competitors to R,2 but
rather as projects sharing a common goal who should support each other whenever possible. For
this reason, gretl provides a way to interact with R and thus enable users to pool the capabilities of
the two packages.
In this chapter, we will explain how to exploit R’s power from within gretl. We assume that the
reader has a working installation of R available and a basic grasp of R’s syntax.3
Despite several valiant attempts, no graphical shell has gained wide acceptance in the R community:
by and large, the standard method of working with R is by writing scripts, or by typing commands
at the R prompt, much in the same way as one would write gretl scripts or work with the gretl
console. In this chapter, the focus will be on the methods available to execute R commands without
leaving gretl.
1 R’shomepage is at http://www.r-project.org/.
2 OK, who are we kidding? But it’s friendly competition!
3 The main reference for R documentation is http://cran.r-project.org/manuals.html. In addition, R tutorials
364
Chapter 39. Gretl and R 365
We will now replicate the above results using R. Select the menu item “Tools, Start GNU R”. A
window similar to the one shown in figure 39.1 should appear.
The actual look of the R window may be somewhat different from what you see in Figure 39.1
(especially for Windows users), but this is immaterial. The important point is that you have a
window where you can type commands to R. If the above procedure doesn’t work and no R window
opens, it means that gretl was unable to launch R. You should ensure that R is installed and working
on your system and that gretl knows where it is. The relevant settings can be found by selecting
the “Tools, Preferences, General” menu entry, under the “Programs” tab.
Assuming R was launched successfully, you will see notification that the data from gretl are avail-
able. In the background, gretl has arranged for two R commands to be executed, one to load the
gretl dataset in the form of a data frame (one of several forms in which R can store data) and one
to attach the data so that the variable names defined in the gretl workspace are available as valid
identifiers within R.
In order to replicate gretl’s OLS estimation, go into the R window and type at the prompt
You should see something similar to Figure 39.2. Surprise — the estimates coincide! To get out, just
close the R window or type q() at the R prompt.
open bjg
arima 0 1 1 ; 0 1 1 ; lg --nc
Table 39.2: Airline model from Box and Jenkins (1976) – selected portion of gretl’s estimates
If we now open an R session as described in the previous subsection, the data-passing mechanism
is slightly different. Since our data were defined in gretl as time series, we use an R time-series
object (ts for short) for the transfer. In this way we can retain in R useful information such as the
periodicity of the data and the sample limits. The downside is that the names of individual series,
as defined in gretl, are not valid identifiers. In order to extract the variable lg, one needs to use the
syntax lg <- gretldata[, "lg"].
ARIMA estimation can be carried out by issuing the following two R commands:
which yield
Coefficients:
ma1 sma1
-0.4018 -0.5569
s.e. 0.0896 0.0731
In either case, you are presented with a window very similar to the editor window used for ordinary
gretl scripts, as in Figure 39.3.
There are two main differences. First, you get syntax highlighting for R’s syntax instead of gretl’s.
Second, clicking on the Execute button (the gears icon), launches an instance of R in which your
commands are executed. Before R is actually run, you are asked if you want to run R interactively
or not (see Figure 39.4).
An interactive run opens an R instance similar to the one seen in the previous section: your data
will be pre-loaded (if the “pre-load data” box is checked) and your commands will be executed.
Once this is done, you will find yourself at the R prompt, where you can enter more commands.
Chapter 39. Gretl and R 368
A non-interactive run, on the other hand, will execute your script, collect the output from R and
present it to you in an output window; R will be run in the background. If, for example, the script
in Figure 39.3 is run non-interactively, a window similar to Figure 39.5 will appear.
and stores it into the file mymatfile.mat in the user’s “dotdir”(see section 14.2). Note that writing
to this special directory, which is sure to exist and be writable by the user, is mandated by the
non-zero value for the third, optional argument to mwrite.
matrix A = mshape(seq(3,14),4,3)
err = mwrite(A, "mymatfile.mat", 1)
A <- gretl.loadmat("mymatfile.mat")
The function gretl.loadmat, which is predefined when R is called from gretl, retrieves the matrix
from dotdir. (The “.mat” extension for gretl matrix files is not compulsory; you can name these
files as you wish.)
It’s also possible to take more control over the details of the transer if you wish. You have the
built-in string variable $dotdir in gretl, while in R you have the same variable under the name
gretl.dotdir. To use a location other than $dotdir you may (a) omit the third argument to
mwrite and supply a full path to the matrix file, and (b) use a more generic approach to reading
the file in R. Here’s an example:
Gretl side:
mwrite(A, "/path/to/mymatfile.mat")
R side:
As an example, we take the airline data and use them to estimate a structural time series model à
la Harvey (1989).4 The model we will use is the Basic Structural Model (BSM), in which a time series
is decomposed into three terms:
yt = µt + γt + εt
where µt is a trend component, γt is a seasonal component and εt is a noise term. In turn, the
following is assumed to hold:
∆µt = βt−1 + ηt
∆βt = ζt
∆s γt = ∆ωt
where ∆s is the seasonal differencing operator, (1 − Ls ), and ηt , ζt and ωt are mutually uncorre-
lated white noise processes. The object of the analysis is to estimate the variances of the noise
components (which may be zero) and to recover estimates of the latent processes µt (the “level”),
βt (the “slope”) and γt .
We will use R’s StructTS command and import the results back into gretl. Once the bjg dataset is
loaded in gretl, we pass the data to R and execute the following script:
However, we are now able to pull the results back into gretl by executing the following commands,
either from the console or by creating a small script:
The first command reads the estimated time-series components from a CSV file, which is the format
that the passing mechanism employs for series. The matrix vars is read from the file vars.mat.
After the above commands have been executed, three new series will have appeared in the gretl
workspace, namely the estimates of the three components; by plotting them together with the
original data, you should get a graph similar to Figure 39.6. The estimates of the variances can be
seen by printing the vars matrix, as in
? print vars
vars (4 x 1)
4 The function package StucTiSM is available to handle this class of models natively in gretl.
Chapter 39. Gretl and R 371
lg level
6.6 6.2
6.4 6
6.2
5.8
6
5.8 5.6
5.6 5.4
5.4 5.2
5.2
5
5
4.8 4.8
4.6 4.6
1949 1955 1961 1949 1955 1961
slope sea
0.01025 0.3
0.25
0.0102 0.2
0.15
0.01015 0.1
0.05
0
0.0101 -0.05
-0.1
0.01005 -0.15
-0.2
0.01 -0.25
1949 1955 1961 1949 1955 1961
0.00077185
0.0000
0.0013969
0.0000
That is,
σ̂η2 = 0.00077185, σ̂ζ2 = 0, 2
σ̂ω = 0.0013969, σ̂ε2 = 0
Notice that, since σ̂ζ2 = 0, the estimate for βt is constant and the level component is simply a
random walk with a drift.
and achieves the same effect as submitting the enclosed R commands via the GUI in the non-
interactive mode (see section 39.3 above). The --send-data option arranges for auto-loading of
the data present in the gretl session, or a subset thereof specified via a named list. The --quiet
option prevents the output from R from being echoed in the gretl output.
Using this method, replicating the example in the previous subsection is rather easy: basically, all it
takes is encapsulating the content of the R script in a foreign. . . end foreign block; see example
39.1.
Chapter 39. Gretl and R 372
open bjg.gdt
append @dotdir/compon.csv
rename level lg_level
rename slope lg_slope
rename sea lg_seas
vars = mread("vars.mat", 1)
append @dotdir/compon.csv
rename level @sx_level
rename slope @sx_slope
rename sea @sx_seas
open bjg.gdt
list X = RStructTS(lg)
Chapter 39. Gretl and R 373
The above syntax, despite being already quite useful by itself, shows its full power when it is used
in conjunction with user-written functions. Listing 39.2 shows how to define a gretl function that
calls R internally.
set R_functions on
foreign language=R
plus_one <- function(q) {
z = q+1
invisible(z)
}
end foreign
scalar b=R.plus_one(2)
The R function plus_one is obviously trivial in itself, but the example shows a couple of points.
First, for this mechanism to work you need to enable R_functions via the set command. Second,
to avoid collision with the gretl function namespace, calls to functions defined in this way must be
prefixed with “R.”, as in R.plus_one. But note, this mechanism will not work if you have defined
a gretl bundle named R: in that case identifiers beginning with “R.” will be understood as referring
to members of the bundle in question.
Built-in R functions may also be called in this way, once R_functions is set on. For example one
can invoke R’s choose function, which computes binomial coefficients:
set R_functions on
scalar b = R.choose(10,4)
Note, however, that the possibilities for use of built-in R functions are limited; only functions
whose arguments and return values are sufficiently generic (basically scalars, matrices or strings)
will work.
Chapter 40
Gretl and Ox
40.1 Introduction
Ox, written by Jurgen A. Doornik (see Doornik, 2007), is described by its author as “an object-
oriented statistical system. At its core is a powerful matrix language, which is complemented by
a comprehensive statistical library. Among the special features of Ox are its speed [and] well-
designed syntax. . . . Ox comes in two versions: Ox Professional and Ox Console. Ox is available for
Windows, Linux, Mac (OS X), and several Unix platforms.” (www.doornik.com)
Ox is proprietary, closed-source software. The command-line version of the program is, however,
available free of change for academic users. Quoting again from Doornik’s website: “The Console
(command line) versions may be used freely for academic research and teaching purposes only. . . .
The Ox syntax is public, and, of course, you may do with your own Ox code whatever you wish.”
If you wish to use Ox in conjunction with gretl please refer to doornik.com for further details on
licensing.
As the reader will no doubt have noticed, most other software that we discuss in this Guide is open-
source and freely available for all users. We make an exception for Ox on the grounds that it is
indeed fast and well designed, and that its statistical library — along with various add-on packages
that are also available—has exceptional coverage of cutting-edge techniques in econometrics. The
gretl authors have used Ox for benchmarking some of gretl’s more advanced features such as
dynamic panel models and state space models.1
+ To enable support for Ox, go to the Tools/Preferences/General menu item and look under the Programs
tab. Find the entry for the path to the oxl executable, that is, the program that runs Ox files (on MS Windows
it is called oxl.exe). Adjust the path if it’s not already right for your system and you should be ready to go.
With support enabled, you can open and edit Ox programs in the gretl GUI. Clicking the “execute”
icon in the editor window will send your code to Ox for execution. Figures 40.1 and Figure 40.2
show an Ox program and part of its output.
In addition you can embed Ox code within a gretl script using a foreign block, as described in
connection with R. A trivial example, which simply prints the gretl data matrix within Ox, is shown
below:
open data4-1
matrix m = { dataset }
mwrite(m, "gretl.mat", 1)
foreign language=Ox
#include <oxstd.h>
main()
{
1 For a review of Ox, see Cribari-Neto and Zarkos (2003) and for a (somewhat dated) comparison of Ox with other
374
Chapter 40. Gretl and Ox 375
The above example illustrates how a matrix can be passed from gretl to Ox. We use the mwrite
function to write a matrix into the user’s “dotdir” (see section 14.2), then in Ox we use the function
gretl_loadmat to retrieve the matrix.
How does gretl_loadmat come to be defined? When gretl writes out the Ox program correspond-
ing to your foreign block it does two things in addition. First, it writes a small utility file named
gretl_io.ox into your dotdir. This contains a definition for gretl_loadmat and also for the
function gretl_export (see below). Second, gretl interpolates into your Ox code a line which in-
cludes this utility file (it is inserted right after the inclusion of oxstd.h, which is needed in all Ox
programs). Note that gretl_loadmat expects to find the named file in the user’s dotdir.
Listing 40.1: Estimation of dynamic panel data model via gretl and Ox
open abdata.gdt
main ()
{
decl dpd = new DPD();
dpd.Load("@dotdir/abdata.csv");
dpd.SetYear("YEAR");
delete dpd;
}
end foreign
41.1 Introduction
GNU Octave, written by John W. Eaton and others, is described as “a high-level language, primar-
ily intended for numerical computations.” The program is oriented towards “solving linear and
nonlinear problems numerically” and “performing other numerical experiments using a language
that is mostly compatible with Matlab.” (www.gnu.org/software/octave) Octave is available in
source-code form (naturally, for GNU software) and also in the form of binary packages for MS Win-
dows and Mac OS X. Numerous contributed packages that extend Octave’s functionality in various
ways can be found at octave.sf.net.
open data4-1
matrix m = { dataset }
mwrite(m, "gretl.mat", 1)
foreign language=Octave
gmat = gretl_loadmat("gretl.mat")
end foreign
We use the mwrite function to write a matrix into the user’s “dotdir” (see section 14.2), then in Oc-
tave we use the function gretl_loadmat to retrieve the matrix. The “magic” behind gretl_loadmat
works in essentially the same way as for Ox (chapter 40).
378
Chapter 41. Gretl and Octave 379
In addition we need some specialized MATLAB files made available by Mario Forni of the Univer-
sity of Modena, at http://morgana.unimore.it/forni_mario/matlab.htm. The files needed are
coheren2.m, coheren.m, coher.m, cospec.m, crosscov.m, crosspec.m, crosspe.m and spec.m.
These are in a form appropriate for MS Windows. On Linux you could run the following shell script
to get the files and remove the Windows end-of-file character (which prevents the files from running
under Octave):
SITE=http://morgana.unimore.it/forni_mario/MYPROG
# download files and delete trailing Ctrl-Z
for f in \
coheren2.m \
coheren.m \
coher.m \
cospec.m \
crosscov.m \
crosspec.m \
crosspe.m \
spec.m ; do
wget $SITE/$f && \
cat $f | tr -d \\032 > tmp.m && mv tmp.m $f
done
The Forni files should be placed in some appropriate directory, and you should tell Octave where
to find them by adding that directory to Octave’s loadpath. On Linux this can be done via an entry
in one’s ~/.octaverc file. For example
addpath("~/stats/octave/forni");
Alternatively, an addpath directive can be written into the Octave script that calls on these files.
With everything set up on the Octave side we now write a gretl script (see Listing 41.1) which opens
a time-series dataset, constructs and writes a matrix containing two series, and defines a foreign
block containing the Octave statements needed to produce the spectral coherence matrix. The
matrix is exported via the gretl_export function, which is automatically defined for you; this
function takes two arguments, a matrix and a file name. The file is written into the user’s “dotdir”,
from where it can be picked up using mread. Finally, we produce a graph from the matrix in gretl.
In the script this is sent to the screen; Figure 41.3 shows the same graph in PDF format.
Chapter 41. Gretl and Octave 381
open data9-7
matrix xy = { PRIME, UNEMP }
mwrite(xy, "xy.mat", 1)
foreign language=Octave
# uncomment and modify the following if necessary
# addpath("~/stats/octave/forni");
xy = gretl_loadmat("xy.mat");
x = xy(:,1);
y = xy(:,2);
# note: the last parameter is the Bartlett window size
h = coher(x, y, 8);
gretl_export(h, "h.mat");
end foreign
h = mread("h.mat", 1)
colnames(h, "coherence")
gnuplot 1 --time-series --with-lines --matrix=h --output=display
0.4
0.3
0.2
0.1
coherence
-0.1
-0.2
-0.3
-0.4
-0.5
0 10 20 30 40 50 60
Stata (www.stata.com) is closed-source, proprietary (and expensive) software and as such is not a
natural companion to gretl. Nonetheless, given Stata’s popularity it is desirable to have a convenient
way of comparing results across the two programs, and to that end we provide some support for
Stata code under the foreign command.
+ To enable support for Stata, go to the Tools/Preferences/General menu item and look under the Programs
tab. Find the entry for the path to the Stata executable. Adjust the path if it’s not already right for your
system and you should be ready to go.
The following example illustrates what’s available. You can send the current gretl dataset to Stata
using the --send-data flag. And having defined a matrix within Stata you can export it for use with
gretl via the gretl_export command: this takes two arguments, the name of the matrix to export
and the filename to use; the file is written to the user’s “dotdir”, from where it can be retrieved
using the mread() function.1 To suppress printed output from Stata you can add the --quiet flag
to the foreign block.
open data4-1
ols 1 0 2 3 --cluster=bedrms
matrix se = $stderr
In addition you can edit “pure” Stata scripts in the gretl GUI and send them for execution as with
native gretl scripts.
Note that Stata coerces all variable names to lower-case on data input, so even if series names in
1 We do not currently offer the complementary functionality of gretl_loadmat, which enables reading of matrices
written by gretl’s mwrite() function in Ox and Octave. This is not at all easy to implement in Stata code.
382
Chapter 42. Gretl and Stata 383
gretl are upper-case, or of mixed case, it’s necessary to use all lower-case in Stata. Also note that
when opening a data file within Stata via the use command it will be necessary to provide the full
path to the file.
Chapter 43
43.1 Introduction
According to www.python.org, Python is “an easy to learn, powerful programming language. It has
efficient high-level data structures and a simple but effective approach to object-oriented program-
ming. Python’s elegant syntax and dynamic typing, together with its interpreted nature, make it an
ideal language for scripting and rapid application development in many areas on most platforms.”
Indeed, Python is widely used in a great variety of contexts. Numerous add-on modules are avail-
able; the ones likely to be of greatest interest to econometricians include NumPy (“the fundamen-
tal package for scientific computing with Python” — see www.numpy.org); SciPy (which builds on
NumPy —see www.scipy.org); and Statsmodels (http://statsmodels.sourceforge.net/).
gretl_dotdir
gretl_loadmat(filename, autodot=1)
gretl_export(M, filename, autodot=1)
The variable gretl_dotdir holds the path to the user’s “dot directory.” The first function loads a
matrix of the given filename as written by gretl’s mwrite function, and the second writes matrix
M, under the given filename, in the format wanted by gretl.
By default the traffic in matrices goes via the dot directory on the Python side; that is, the name of
this directory is prepended to filename for both reading and writing. (This is complementary to
use of the export and import parameters with gretl’s mwrite and mread functions, respectively.)
However, if you wish to take control over the reading and writing locations you can supply a zero
value for autodot when calling gretl_loadmat and gretl_export: in that case the filename
argument is used as is.
Note that gretl_loadmat and gretl_export depend on NumPy; they make use of the functions
loadtxt and savetxt respectively. Nonetheless, the presence of NumPy is not an absolute require-
ment if you don’t need to use these two functions.
384
Chapter 43. Gretl and Python 385
the NIST-certified values as a benchmark;1 the error values correspond to the number of correct
digits (with a maximum of 15). The results will likely differ somewhat by computer architecture
and compiler.
open longley.gdt -q
list LX = prdefl .. year
ols employ 0 LX -q
matrix b_gretl = $coeff
foreign language=python
import numpy as np
X = gretl_loadmat(’alldata.mat’, 1)
# NumPy’s OLS
b = np.linalg.lstsq(X[:,1:], X[:,0])[0]
gretl_export(np.transpose(np.matrix(b)), ’py_b.mat’, 1)
end foreign
gretl python
12.844 12.850
11.528 11.414
12.393 12.401
13.135 13.121
13.738 13.318
12.587 12.363
12.848 12.852
Column means
12.725 12.617
1 See http://www.itl.nist.gov/div898/strd/lls/data/Longley.shtml.
Chapter 44
44.1 Introduction
According to julialang.org, Julia is “a high-level, high-performance dynamic programming lan-
guage for technical computing, with syntax that is familiar to users of other technical computing
environments. It provides a sophisticated compiler, distributed parallel execution, numerical ac-
curacy, and an extensive mathematical function library.” Julia is well known for being very fast;
however, you should be aware that by default starting Julia takes some time due to Just-in-Time
compilation of the input. This fixed cost is well worth bearing if you are asking Julia to perform a
big computation, but small jobs are likely to run faster if you use the (Julia-specific) --no-compile
option with the foreign command.
gretl_dotdir
gretl_loadmat(filename, autodot=true)
gretl_export(M, filename, autodot=true)
The variable gretl_dotdir holds the path to the user’s “dot directory.” The first function loads a
matrix of the given filename as written by gretl’s mwrite function, and the second writes matrix
M, under the given filename, in the format wanted by gretl.
By default the traffic in matrices goes via the dot directory on the Julia side; that is, the name of
this directory is prepended to filename for both reading and writing. (This is complementary to
use of the export and import parameters with gretl’s mwrite and mread functions, respectively.)
However, if you wish to take control over the reading and writing locations you can supply a zero
value for autodot when calling gretl_loadmat and gretl_export: in that case the filename
argument is used as is.
44.3 Illustration
Listing 44.1 shows a minimal example of how to interact with Julia from a gretl script.
Since this is a very small job JIT compilation is not worthwhile; in our testing the script runs almost
4 times faster if the Julia block is opened with
This has the effect of passing the option --compile=no to the Julia executable.
386
Chapter 44. Gretl and Julia 387
check (4 x 4)
Troubleshooting gretl
388
Chapter 45. Troubleshooting gretl 389
When you start gretl in this mode, a “console window” appears as well as the gretl window, and
stderr output goes to the console. To copy this output, click at the top left of the console window
for a menu (Figure 45.2): first do “Select all”, then “Copy.” You can paste the results into Notepad
or similar.
If the program fails to start from the prompt, it’s not a gretl issue but rather that the program’s
home directory is not in your path, or the program is not installed (properly). For details on
modifying your path please see the documentation or online help for your operating system or
shell.
1 The exception to this rule is the invocation of gnuplot under MS Windows, where a full path to the program is given.
Chapter 46
The gretl package includes the command-line program gretlcli. On Linux it can be run from a
terminal window (xterm, rxvt, or similar), or at the text console. Under MS Windows it can be run in
a console window (sometimes inaccurately called a “DOS box”). gretlcli has its own help file, which
may be accessed by typing “help” at the prompt. It can be run in batch mode, sending output
directly to a file (see also the Gretl Command Reference).
If gretlcli is linked to the readline library (this is automatically the case in the MS Windows version;
also see Appendix C), the command line is recallable and editable, and offers command completion.
You can use the Up and Down arrow keys to cycle through previously typed commands. On a given
command line, you can use the arrow keys to move around, in conjunction with Emacs editing
keystrokes.1 The most common of these are:
Keystroke Effect
Ctrl-a go to start of line
Ctrl-e go to end of line
Ctrl-d delete character to right
where “Ctrl-a” means press the “a” key while the “Ctrl” key is also depressed. Thus if you want
to change something at the beginning of a command, you don’t have to backspace over the whole
line, erasing as you go. Just hop to the start and add or delete characters. If you type the first
letters of a command name then press the Tab key, readline will attempt to complete the command
name for you. If there’s a unique completion it will be put in place automatically. If there’s more
than one completion, pressing Tab a second time brings up a list.
Probably the most useful mode for heavy-duty work with gretlcli is batch (non-interactive) mode,
in which the program reads and processes a script, and sends the output to file. For example
Note that scriptfile is treated as a program argument; only the output file requires redirection (>).
Don’t forget the -b (batch) switch, otherwise the program will wait for user input after executing
the script (and if output is redirected, the program will appear to “hang”).
1 Actually, the key bindings shown below are only the defaults; they can be customized. See the readline manual.
390
Part IV
Appendices
391
Appendix A
• data.xml is an XML file conforming to the gretldata DTD mentioned above, holding all the
metadata.
Binary values are saved in the endianness of the machine on which they’re written; the header
information enables gretl to convert to the endianness of the host on which the data are read if
need be.
The rationale for introducing the binary gdtb format is that for very large datasets it is a lot faster
to write and read data in this form rather than as text. For small to moderately sized datasets
(say, up to 10 megabytes or so) there is no appreciable advantage in the binary format and we
recommend use of plain gdt.
Some illustrative timings are shown in Table A.1; these were obtained on a Lenovo ThinkPad X1 Car-
bon running gretl 1.9.15. The datasets contained a mixture of random normal series and random
binary (dummy) series. The largest comprised 50000 observations on 1000 series and the smallest
5000 observations on 250 series. As can be seen, there is a big time saving from the binary for-
mat when writing (and to a lesser extent, when reading) a dataset in the hundreds of megabytes
392
Appendix A. Data file details 393
range. At a size of around 10 megabytes, however, gdt files can be both written and read in under
a second, surely fast enough for most purposes.
381 MB 38 MB 19 MB 10 MB
format write read write read write read write read
gdt 21.76 4.91 2.20 0.48 1.19 0.32 0.59 0.16
gdtb 3.82 1.10 0.39 0.11 0.25 0.06 0.13 0.03
Table A.1: Data write and read timings in seconds for datasets of various sizes in megabytes. The MB
numbers represent the size of the datasets in memory; files of both formats are substantially smaller when
compression is applied.
The first field is the series name. The second is a description of the series (maximum 128 charac-
ters). On the second line the first field is a frequency code: M for monthly, Q for quarterly, A for
annual, B for business-daily (daily with five days per week), D for 7-day daily, S for 6-day daily, U
for undated.
No other frequencies are accepted at present. Then comes the starting date (N.B. with two digits
following the point for monthly data, one for quarterly data, none for annual), a space, a hyphen,
another space, the ending date, the string “n = ” and the integer number of observations. In the
case of daily data the starting and ending dates should be given in the form YYYY-MM-DD. This
format must be respected exactly.
Optionally, the first line of the index file may contain a short comment (up to 64 characters) on the
source and nature of the data, following a hash mark. For example:
The corresponding binary database file holds the data values, represented as “floats”, that is, single-
precision floating-point numbers, typically taking four bytes apiece. The numbers are packed “by
variable”, so that the first n numbers are the observations of variable 1, the next m the observations
on variable 2, and so on.
Appendix B
Since version 1.7.5, gretl provides a method for retrieving data from databases which support the
Open Database Connectivity (ODBC) standard. Most users won’t be interested in this, but there may
be some for whom this feature matters a lot — typically, those who work in an environment where
huge data collections are accessible via a Data Base Management System (DBMS).
In the following section we explain what is needed for ODBC support in gretl. We provide some
background information on how ODBC works in section B.2, and explain the details of getting gretl
to retrieve data from a database in section B.3.
ODBC
query
data
For the above mechanism to work, it is necessary that the relevant ODBC software is installed
and working on the client machine (contact your DB administrator for details). At this point, the
database (or databases) that the server provides will be accessible to the client as a data source
with a specific identifier (a Data Source Name or DSN); in most cases, a username and a password
are required to connect to the data source.
Once the connection is established, the user sends a query to ODBC, which contacts the database
manager, collects the results and sends them back to the user. The query is almost invariably
394
Appendix B. Data import via ODBC 395
formulated in a special language used for the purpose, namely SQL.1 We will not provide here an
SQL tutorial: there are many such tutorials on the Net; besides, each database manager tends to
support its own SQL dialect so the precise form of an SQL query may vary slightly if the DBMS on
the other end is Oracle, MySQL, PostgreSQL or something else.
Suffice it to say that the main statement for retrieving data is the SELECT statement. Within a DBMS,
data are organized in tables, which are roughly equivalent to spreadsheets. The SELECT statement
returns a subset of a table, which is itself a table. For example, imagine that the database holds a
table called “NatAccounts”, containing the data shown in Table B.1.
Gretl provides a mechanism for forwarding your query to the DBMS via ODBC and including the
results in your currently open dataset.
B.3 Syntax
At present gretl does not offer a graphical interface for ODBC import; this must be done via the
command line interface. The two commands used for fetching data via an ODBC connection are
open and data.
The open command is used for connecting to a DBMS: its syntax is
The user and password items are optional; the effect of this command is to initiate an ODBC
connection. It is assumed that the machine gretl runs on has a working ODBC client installed.
In order to actually retrieve the data, the data command is used. Its syntax is:
1 See http://en.wikipedia.org/wiki/SQL.
Appendix B. Data import via ODBC 396
where:
series is a list of names of gretl series to contain the incoming data, separated by spaces. Note that
these series need not exist pior to the ODBC import.
format-string is an optional parameter, used to handle cases when a “rectangular” organisation of
the database cannot be assumed (more on this later);
query-string is a string containing the SQL statement used to extract the data.2
There should be no spaces around the equals signs in the obs-format and query fields in the data
command.
The query-string can, in principle, contain any valid SQL statement which results in a table. This
string may be specified directly within the command, as in
which will store into the gretl variable x the content of the column foo from the table bar. However,
since in a real-life situation the string containing the SQL statement may be rather long, it may be
best to store it in a string variable. For example:
(The series named index is automatically added to a dataset created via the nulldata command.)
The format specifiers available for use with obs-format are as follows:
could be imported at a time. This variant is still accepted for the sake of backward compatibility.
Appendix B. Data import via ODBC 397
In addition the format can include literal characters to be passed through, such as slashes or colons,
to make the resulting string compatible with gretl’s observation identifiers.
For example, consider the following fictitious case: we have a 5-days-per-week dataset, to which we
want to add the stock index for the Verdurian market;3 it so happens that in Verduria Saturdays
are working days but Wednesdays are not. We want a column which does not contain data on
Saturdays, because we wouldn’t know where to put them, but at the same time we want to place
missing values on all the Wednesdays.
In this case, the following syntax could be used
The column VerdSE holds the data to be fetched, which will go into the gretl series y. The first
three columns are used to construct a string which identifies the day. Daily dates take the form
YYYY/MM/DD in gretl. If a row from the DBMS produces the observation string 2008/04/01 this will
match OK (it’s a Tuesday), but 2008/04/05 will not match since it is a Saturday; the corresponding
row will therefore be discarded. On the other hand, since no string 2008/04/23 will be found in
the data coming from the DBMS (it’s a Wednesday), that entry is left blank in our series y.
B.4 Examples
In the following examples, we will assume that access is available to a database known to ODBC
with the data source name “AWM”, with username “Otto” and password “Bingo”. The database
“AWM” contains quarterly data in two tables (see B.3 and B.4):
3 See http://www.almeopedia.com/index.php/Verduria.
Appendix B. Data import via ODBC 398
The table Consump is the classic “rectangular” dataset; that is, its internal organization is the same
as in a spreadsheet or econometrics package: each row is a data point and each column is a variable.
The structure of the DATA table is different: each record is one figure, stored in the column xval,
and the other fields keep track of which variable it belongs to, for which date.
nulldata 160
setobs 4 1970:1 --time
open dsn=AWM user=Otto password=Bingo --odbc
Listing B.1 shows a query for two series: first we set up an empty quarterly dataset. Then we
connect to the database using the open statement. Once the connection is established we retrieve
two columns from the Consump table. No observation string is required because the data already
have a suitable structure; we need only import the relevant columns.
In example B.2, by contrast, we make use of the observation string since we are drawing from the
DATA table, which is not rectangular. The SQL statement stored in the string S produces a table with
three columns. The ORDER BY clause ensures that the rows will be in chronological order, although
this is not strictly necessary in this case.
Appendix B. Data import via ODBC 399
Listing B.3 shows what happens if the rows in the outcome from the SELECT statement do not
match the observations in the currently open gretl dataset. The query includes a condition which
filters out all the data from the first quarter. The query result (invisible to the user) would be
something like
+------+------+---------------+
| year | qtr | xval |
+------+------+---------------+
| 1970 | 2 | 7.8705000000 |
| 1970 | 3 | 7.5600000000 |
| 1970 | 4 | 7.1892000000 |
| 1971 | 2 | 5.8679000000 |
| 1971 | 3 | 6.2442000000 |
| 1971 | 4 | 5.9811000000 |
| 1972 | 2 | 4.6883000000 |
| 1972 | 3 | 4.6302000000 |
...
Internally, gretl fills the variable bar with the corresponding value if it finds a match; otherwise, NA
is used. Printing out the variable bar thus produces
Obs bar
1970:1
1970:2 7.8705
1970:3 7.5600
1970:4 7.1892
1971:1
1971:2 5.8679
1971:3 6.2442
1971:4 5.9811
1972:1
1972:2 4.6883
1972:3 4.6302
...
Appendix C
Building gretl
Here we give instructions detailed enough to allow a user with only a basic knowledge of a Unix-
type system to build gretl. These steps were tested on a fresh installation of Debian Etch. For other
Linux distributions (especially Debian-based ones, like Ubuntu and its derivatives) little should
change. Other Unix-like operating systems such as Mac OS X and BSD would probably require more
substantial adjustments.
In this guided example, we will build gretl complete with documentation. This introduces a few
more requirements, but gives you the ability to modify the documentation files as well, like the
help files or the manuals.
• Gnuplot
• ImageMagick
We also assume that the user has administrative privileges and knows how to install packages. The
examples below are carried out using the apt-get shell command, but they can be performed with
menu-based utilities like aptitude, dselect or the GUI-based program synaptic. Users of Linux
distributions which employ rpm packages (e.g. Red Hat/Fedora, Mandriva, SuSE) may want to refer
to the dependencies page on the gretl website.
The first step is installing the C compiler and related basic utilities, if these are not already in place.
On a Debian (or derivative) system, these are contained in a bunch of packages that can be installed
via the command
Then it is necessary to install the “development” (dev) packages for the libraries that gretl uses:
400
Appendix C. Building gretl 401
Library command
GLIB apt-get install libglib2.0-dev
GTK 3.0 apt-get install libgtk3.0-dev
PNG apt-get install libpng12-dev
XSLT apt-get install libxslt1-dev
LAPACK apt-get install liblapack-dev
FFTW apt-get install libfftw3-dev
READLINE apt-get install libreadline-dev
ZLIB apt-get install zlib1g-dev
XML apt-get install libxml2-dev
GMP apt-get install libgmp-dev
CURL apt-get install libcurl4-gnutls-dev
MPFR apt-get install libmpfr-dev
It is possible to substitute GTK 2.0 for GTK 3.0. The dev packages for these libraries are necessary
to compile gretl—you’ll also need the plain, non-dev library packages to run gretl, but most of
these should already be part of a standard installation. In order to enable other optional features,
like audio support, you may need to install more libraries.
+ The above steps can be much simplified on Linux systems that provide deb-based package managers,
such as Debian and its derivatives (Ubuntu, Knoppix and other distributions). The command
apt-get build-dep gretl
will download and install all the necessary packages for building the version of gretl that is currently present
in your APT sources. Techincally, this does not guarantee that all the software necessary to build the git
version is included, because the version of gretl on your repository may be quite old and build requirements
may have changed in the meantime. However, the chances of a mismatch are rather remote for a reasonably
up-to-date system, so in most cases the above command should take care of everything correctly.
2. Unzip and untar the package. On a system with the GNU utilities available, the command
would be tar xvfJ gretl-N.tar.xz (replace N with the specific version number of the file
you downloaded at step 1).
3. Change directory to the gretl source directory created at step 2 (e.g. gretl-1.10.2).
To work with git you’ll first need to install the git client program if it’s not already on your sys-
tem. Relevant resources you may wish to consult include the main git website at git-scm.com and
instructions specific to gretl: gretl git basics.
Appendix C. Building gretl 402
When grabbing the git sources for the first time, you should first decide where you want to store
the code. For example, you might create a directory called git under your home directory. Open a
terminal window, cd into this directory, and type the following commands:
At this point git should create a subdirectory named gretl-git and fill it with the current sources.
When you want to update the source, this is very simple: just move into the gretl-git directory
and type
git pull
Assuming you’re now in the gretl-git directory, you can proceed in the same manner as with the
released source package.
./configure --help
first to see what options are available. One option you way wish to tweak is --prefix. By default
the installation goes under /usr/local but you can change this. For example
./configure --prefix=/usr
--enable-openmp
By default the gretl GUI is built using version 3.0 of the GTK library, if available, otherwise version
2.0. If you have both versions installed and prefer to use GTK 2.0, use the flag
--enable-gtk2
In order to have the documentation built, we need to pass the relevant option to configure, as in
--enable-build-doc
But please note that this option will work only if you are using the git source.
+ In order to build the documentation, there is the possibility that you will have to install some extra software
on top of the packages mentioned in the previous section. For example, you may need some extra LATEX
packages to compile the manuals. Two of the required packages, that not every standard LATEX installation
include, are typically pifont.sty and appendix.sty. You could install the corresponding packages from your
distribution or you could simply download them from CTAN and install them by hand.
This, for example, if you want to install under /usr, with OpenMP support, and also build the
documentation, you would do
Appendix C. Building gretl 403
./configure --prefix=/usr \
--enable-openmp \
--enable-build-doc
You will see a number of checks being run, and if everything goes according to plan, you should
see a summary similar to that displayed in Listing C.1.
Configuration:
+ If you’re using git, it’s a good idea to re-run the configure script after doing an update. This is not always
necessary, but sometimes it is, and it never does any harm. For this purpose, you may want to write a little
shell script that calls configure with any options you want to use.
make
This step will likely take several minutes to complete; a lot of output will be produced on screen.
Once this is done, you can install your freshly baked copy of gretl on your system via
make install
Appendix C. Building gretl 404
On most systems, the make install command requires you to have administrative privileges.
Hence, either you log in as root before launching make install or you may want to use the sudo
utility, as in:
Now try if everything works: go back to your home directory and run gretl
cd ~
gretl &
If all is well, you ought to see gretl start, at which point just exit the program in the usual way. On
the other hand, there is the possibility that gretl doesn’t start and instead you see a message like
sudo ldconfig
Numerical accuracy
Gretl uses double-precision arithmetic throughout — except for the multiple-precision plugin in-
voked by the menu item “Model, Other linear models, High precision OLS” which represents floating-
point values using a number of bits given by the environment variable GRETL_MP_BITS (default
value 256).
The normal equations of Least Squares are by default solved via Cholesky decomposition, which
is highly accurate provided the matrix of cross-products of the regressors, X 0 X, is not very ill
conditioned. If this problem is detected, gretl automatically switches to use QR decomposition.
The program has been tested rather thoroughly on the statistical reference datasets provided by
NIST (the U.S. National Institute of Standards and Technology) and a full account of the results may
be found on the gretl website (follow the link “Numerical accuracy”).
To date, two published reviews have discussed gretl’s accuracy: Giovanni Baiocchi and Walter Dis-
taso (2003), and Talha Yalta and Yasemin Yalta (2007). We are grateful to these authors for their
careful examination of the program. Their comments have prompted several modifications includ-
ing the use of Stephen Moshier’s cephes code for computing p-values and other quantities relating
to probability distributions (see netlib.org), changes to the formatting of regression output to en-
sure that the program displays a consistent number of significant digits, and attention to compiler
issues in producing the MS Windows version of gretl (which at one time was slighly less accurate
than the Linux version).
Gretl now includes a “plugin” that runs the NIST linear regression test suite. You can find this
under the “Tools” menu in the main window. When you run this test, the introductory text explains
the expected result. If you run this test and see anything other than the expected result, please
send a bug report to cottrell@wfu.edu.
All regression statistics are printed to 6 significant figures in the current version of gretl (except
when the multiple-precision plugin is used, in which case results are given to 12 figures). If you
want to examine a particular value more closely, first save it (for example, using the genr command)
then print it using printf, to as many digits as you like (see the Gretl Command Reference).
405
Appendix E
Gretl’s capabilities are substantial, and are expanding. Nonetheless you may find there are some
things you can’t do in gretl, or you may wish to compare results with other programs. If you are
looking for complementary functionality in the realm of free, open-source software we recommend
the following programs. The self-description of each program is taken from its website.
• GNU Octave www.octave.org: “GNU Octave is a high-level language, primarily intended for
numerical computations. It provides a convenient command line interface for solving linear
and nonlinear problems numerically, and for performing other numerical experiments using
a language that is mostly compatible with Matlab. It may also be used as a batch-oriented
language.”
• JMulTi www.jmulti.de: “JMulTi was originally designed as a tool for certain econometric pro-
cedures in time series analysis that are especially difficult to use and that are not available
in other packages, like Impulse Response Analysis with bootstrapped confidence intervals for
VAR/VEC modelling. Now many other features have been integrated as well to make it possi-
ble to convey a comprehensive analysis.” Comment: JMulTi is a java GUI program: you need
a java run-time environment to make use of it.
As mentioned above, gretl offers the facility of exporting data in the formats of both Octave and
R. In the case of Octave, the gretl data set is saved as a single matrix, X. You can pull the X matrix
apart if you wish, once the data are loaded in Octave; see the Octave manual for details. As for R,
the exported data file preserves any time series structure that is apparent to gretl. The series are
saved as individual structures. The data should be brought into R using the source() command.
In addition, gretl has a convenience function for moving data quickly into R. Under gretl’s “Tools”
menu, you will find the entry “Start GNU R”. This writes out an R version of the current gretl
data set (in the user’s gretl directory), and sources it into a new R session. The particular way
R is invoked depends on the internal gretl variable Rcommand, whose value may be set under the
“Tools, Preferences” menu. The default command is RGui.exe under MS Windows. Under X it is
xterm -e R. Please note that at most three space-separated elements in this command string will
be processed; any extra elements are ignored.
406
Appendix F
Listing of URLs
FFTW3 http://www.fftw.org/
JRSoftware http://www.jrsoftware.org/
Minpack http://www.netlib.org/minpack/
407
Bibliography
Akaike, H. (1974) ‘A new look at the statistical model identification’, IEEE Transactions on Auto-
matic Control AC-19: 716–723.
Anderson, T. W. and C. Hsiao (1981) ‘Estimation of dynamic models with error components’, Jour-
nal of the American Statistical Association 76: 598–606.
Arellano, M. and S. Bond (1991) ‘Some tests of specification for panel data: Monte carlo evidence
and an application to employment equations’, The Review of Economic Studies 58: 277–297.
Baiocchi, G. and W. Distaso (2003) ‘GRETL: Econometric software for the GNU generation’, Journal
of Applied Econometrics 18: 105–110.
Baltagi, B. H. and Y.-J. Chang (1994) ‘Incomplete panels: A comparative study of alternative esti-
mators for the unbalanced one-way error component regression model’, Journal of Econometrics
62: 67–89.
Baltagi, B. H. and Q. Li (1990) ‘A lagrange multiplier test for the error components model with
incomplete panels’, Econometric Reviews 9: 103–107.
Baltagi, B. H. and P. X. Wu (1999) ‘Unequally spaced panel data regressions with AR(1) distur-
bances’, Econometric Theory 15: 814–823.
Baxter, M. and R. G. King (1999) ‘Measuring business cycles: Approximate band-pass filters for
economic time series’, The Review of Economics and Statistics 81(4): 575–593.
Beck, N. and J. N. Katz (1995) ‘What to do (and not to do) with time-series cross-section data’, The
American Political Science Review 89: 634–647.
Bera, A. K., C. M. Jarque and L. F. Lee (1984) ‘Testing the normality assumption in limited depen-
dent variable models’, International Economic Review 25: 563–578.
Berndt, E., B. Hall, R. Hall and J. Hausman (1974) ‘Estimation and inference in nonlinear structural
models’, Annals of Economic and Social Measurement 3(4): 653–665.
Bhargava, A., L. Franzini and W. Narendranathan (1982) ‘Serial correlation and the fixed effects
model’, Review of Economic Studies 49: 533–549.
Blundell, R. and S. Bond (1998) ‘Initial conditions and moment restrictions in dynamic panel data
models’, Journal of Econometrics 87: 115–143.
Bond, S., A. Hoeffler and J. Temple (2001) ‘GMM estimation of empirical growth models’. Economics
Papers from Economics Group, Nuffield College, University of Oxford, No 2001-W21.
408
Bibliography 409
Boswijk, H. P. and J. A. Doornik (2004) ‘Identifying, estimating and testing restricted cointegrated
systems: An overview’, Statistica Neerlandica 58(4): 440–465.
Box, G. E. P. and G. Jenkins (1976) Time Series Analysis: Forecasting and Control, San Franciso:
Holden-Day.
Brand, C. and N. Cassola (2004) ‘A money demand system for euro area M3’, Applied Economics
36(8): 817–838.
Butterworth, S. (1930) ‘On the theory of filter amplifiers’, Experimental Wireless & The Wireless
Engineer 7: 536–541.
Byrd, R. H., P. Lu, J. Nocedal and C. Zhu (1995) ‘A limited memory algorithm for bound constrained
optimization’, SIAM Journal on Scientific Computing 16(5): 1190–1208.
Caselli, F., G. Esquivel and F. Lefort (1996) ‘Reopening the convergence debate: A new look at
cross-country growth empirics’, Journal of Economic Growth 1(3): 363–389.
Chesher, A. and M. Irish (1987) ‘Residual analysis in the grouped and censored normal linear
model’, Journal of Econometrics 34: 33–61.
Choi, I. (2001) ‘Unit root tests for panel data’, Journal of International Money and Finance 20(2):
249–272.
Cleveland, W. S. (1979) ‘Robust locally weighted regression and smoothing scatterplots’, Journal
of the American Statistical Association 74(368): 829–836.
Cottrell, A. (2017) ‘Random effects estimators for unbalanced panel data: a Monte Carlo analysis’.
gretl working papers, number 4. URL https://ideas.repec.org/p/anc/wgretl/4.html.
Cottrell, A. and R. Lucchetti (2016) Gretl Function Package Guide, gretl documentation. URL http:
//sourceforge.net/projects/gretl/files/manual/.
Cribari-Neto, F. and S. G. Zarkos (2003) ‘Econometric and statistical computing using Ox’, Compu-
tational Economics 21: 277–295.
Davidson, R. and E. Flachaire (2001) ‘The wild bootstrap, tamed at last’. GREQAM Document de
Travail 99A32. URL http://russell.vcharite.univ-mrs.fr/GMMboot/wild5-euro.pdf.
Davidson, R. and J. G. MacKinnon (1993) Estimation and Inference in Econometrics, New York:
Oxford University Press.
(2004) Econometric Theory and Methods, New York: Oxford University Press.
Doornik, J. A. (1995) ‘Testing general restrictions on the cointegrating space’. Discussion Paper,
Nuffield College. URL http://www.doornik.com/research/coigen.pdf.
Doornik, J. A., M. Arellano and S. Bond (2006) Panel Data estimation using DPD for Ox.
Bibliography 410
Doornik, J. A. and H. Hansen (1994) ‘An omnibus test for univariate and multivariate normality’.
Working paper, Nuffield College, Oxford.
Durbin, J. and S. J. Koopman (2012) Time Series Analysis by State Space Methods, Oxford: Oxford
University Press, second edn.
Elliott, G., T. J. Rothenberg and J. H. Stock (1996) ‘Efficient tests for an autoregressive unit root’,
Econometrica 64: 813–836.
Engle, R. F. and C. W. J. Granger (1987) ‘Co-integration and error correction: Representation, esti-
mation, and testing’, Econometrica 55: 251–276.
Fiorentini, G., G. Calzolari and L. Panattoni (1996) ‘Analytic derivatives and the computation of
GARCH estimates’, Journal of Applied Econometrics 11: 399–417.
Frigo, M. and S. G. Johnson (2005) ‘The design and implementation of FFTW3’, Proceedings of the
IEEE 93 2: 216–231.
Golub, G. H. and C. F. Van Loan (1996) Matrix Computations, Baltimore and London: The John
Hopkins University Press, third edn.
Goossens, M., F. Mittelbach and A. Samarin (2004) The LATEX Companion, Boston: Addison-Wesley,
second edn.
Gould, W. (2013) ‘Interpreting the intercept in the fixed-effects model’. URL http://www.stata.
com/support/faqs/statistics/intercept-in-fixed-effects-model/.
Gourieroux, C., A. Monfort, E. Renault and A. Trognon (1987) ‘Generalized residuals’, Journal of
Econometrics 34: 5–32.
Greene, W. H. (2000) Econometric Analysis, Upper Saddle River, NJ: Prentice-Hall, fourth edn.
(2003) Econometric Analysis, Upper Saddle River, NJ: Prentice-Hall, fifth edn.
Hamilton, J. D. (1994) Time Series Analysis, Princeton, NJ: Princeton University Press.
Hannan, E. J. and B. G. Quinn (1979) ‘The determination of the order of an autoregression’, Journal
of the Royal Statistical Society, B 41: 190–195.
Hansen, L. P. and T. J. Sargent (2013) Recursive models of dynamic linear economies, Princeton
University Press.
Harvey, A. C. (1989) Forecasting, Structural Time Series Models and the Kalman Filter, Cambridge:
Cambridge University Press.
Harvey, A. C. and A. Jaeger (1993) ‘Detrending, stylized facts and the business cycle’, Journal of
Applied Econometrics 8(3): 231–247.
Heckman, J. (1979) ‘Sample selection bias as a specification error’, Econometrica 47: 153–161.
Hodrick, R. and E. C. Prescott (1997) ‘Postwar U.S. business cycles: An empirical investigation’,
Journal of Money, Credit and Banking 29: 1–16.
Bibliography 411
Im, K. S., M. H. Pesaran and Y. Shin (2003) ‘Testing for unit roots in heterogeneous panels’, Journal
of Econometrics 115: 53–74.
Islam, N. (1995) ‘Growth Empirics: A Panel Data Approach’, The Quarterly Journal of Economics
110(4): 1127–1170.
Kalbfleisch, J. D. and R. L. Prentice (2002) The Statistical Analysis of Failure Time Data, New York:
Wiley, second edn.
Keane, M. P. and K. I. Wolpin (1997) ‘The career decisions of young men’, Journal of Political
Economy 105: 473–522.
King, R. G. and S. T. Rebelo (1993) ‘Low frequency filtering and real business cycles’, Journal of
Economic dynamics and Control 17(1-2): 207–231.
Koenker, R. (1994) ‘Confidence intervals for regression quantiles’. In P. Mandl and M. Huskova
(eds.), Asymptotic Statistics, pp. 349–359. New York: Springer-Verlag.
Koenker, R. and K. Hallock (2001) ‘Quantile regression’, Journal of Economic Perspectives 15(4):
143–156.
Koenker, R. and J. Machado (1999) ‘Goodness of fit and related inference processes for quantile
regression’, Journal of the American Statistical Association 94: 1296–1310.
Koenker, R. and Q. Zhao (1994) ‘L-estimation for linear heteroscedastic models’, Journal of Non-
parametric Statistics 3: 223–235.
Koopman, S. J. (1993) ‘Disturbance smoother for state space models’, Biometrika 80: 117–126.
Koopman, S. J., N. Shephard and J. A. Doornik (1999) ‘Statistical algorithms for models in state
space using SsfPack 2.2’, Econometrics Journal 2: 107–160.
Kwiatkowski, D., P. C. B. Phillips, P. Schmidt and Y. Shin (1992) ‘Testing the null of stationarity
against the alternative of a unit root: How sure are we that economic time series have a unit
root?’, Journal of Econometrics 54: 159–178.
Levin, A., C.-F. Lin and J. Chu (2002) ‘Unit root tests in panel data: asymptotic and finite-sample
properties’, Journal of Econometrics 108: 1–24.
Lucchetti, R., L. Papi and A. Zazzaro (2001) ‘Banks’ inefficiency and economic growth: A micro
macro approach’, Scottish Journal of Political Economy 48: 400–424.
Lütkepohl, H. (2005) New Intoduction to Multiple Time Series Analysis, Berlin: Springer.
MacKinnon, J. G. (1996) ‘Numerical distribution functions for unit root and cointegration tests’,
Journal of Applied Econometrics 11: 601–618.
Magnus, J. R. and H. Neudecker (1988) Matrix Differential Calculus with Applications in Statistics
and Econometrics, John Wiley & Sons.
McAleer, M. and L. Oxley (1999) Practical Issues in Cointegration Analysis, Oxford: Blackwell.
McCullough, B. D. and C. G. Renfro (1998) ‘Benchmarks and software standards: A case study of
GARCH procedures’, Journal of Economic and Social Measurement 25: 59–71.
Melard, G. (1984) ‘Algorithm AS 197: A Fast Algorithm for the Exact Maximum Likelihood of
Autoregressive-Moving Average Models’, Journal of the Royal Statistical Society. Series C (Applied
Statistics 33(1): 104–114.
Bibliography 412
Morales, J. L. and J. Nocedal (2011) ‘Remark on Algorithm 778: L-BFGS-B: Fortran routines for
large-scale bound constrained optimization’, ACM Transactions on Mathematical Software 38(1):
1–4.
Mroz, T. (1987) ‘The sensitivity of an empirical model of married women’s hours of work to eco-
nomic and statistical assumptions’, Econometrica 5: 765–799.
Nadaraya, E. A. (1964) ‘On estimating regression’, Theory of Probability and its Applications 9:
141–142.
Nash, J. C. (1990) Compact Numerical Methods for Computers: Linear Algebra and Function Min-
imisation, Bristol: Adam Hilger, second edn.
Nerlove, M. (1971) ‘Further evidence on the estimation of dynamic economic relations from a time
series of cross sections’, Econometrica 39: 359–382.
Newey, W. K. and K. D. West (1987) ‘A simple, positive semi-definite, heteroskedasticity and auto-
correlation consistent covariance matrix’, Econometrica 55: 703–708.
(1994) ‘Automatic lag selection in covariance matrix estimation’, Review of Economic Stud-
ies 61: 631–653.
Okui, R. (2009) ‘The optimal choice of moments in dynamic panel data models’, Journal of Econo-
metrics 151(1): 1–16.
Pelagatti, M. (2011) ‘State space methods in Ox/SsfPack’, Journal of Statistical Software 41(3): 1–25.
Pollock, D. S. G. (2000) ‘Trend estimation and de-trending via rational square-wave filters’, Journal
of Econometrics 99(2): 317–334.
Portnoy, S. and R. Koenker (1997) ‘The Gaussian hare and the Laplacian tortoise: computability of
squared-error versus absolute-error estimators’, Statistical Science 12(4): 279–300.
Press, W., S. Teukolsky, W. Vetterling and B. Flannery (2007) Numerical Recipes: The Art of Scientific
Computing, Cambridge University Press, 3 edn.
Ramanathan, R. (2002) Introductory Econometrics with Applications, Fort Worth: Harcourt, fifth
edn.
Rao, C. R. (1973) Linear Statistical Inference and its Applications, New York: Wiley, second edn.
Sephton, P. S. (1995) ‘Response surface estimates of the KPSS stationarity test’, Economics Letters
47: 255–261.
Steinhaus, S. (1999) ‘Comparison of mathematical programs for data analysis (edition 3)’. Univer-
sity of Frankfurt. URL http://www.informatik.uni-frankfurt.de/~stst/ncrunch/.
Stock, J. H. and M. W. Watson (1999) ‘Forecasting inflation’, Journal of Monetary Economics 44(2):
293–335.
Bibliography 413
(2008) ‘Heteroskedasticity-robust standard errors for fixed effects panel data regression’,
Econometrica 76(1): 155–174.
Stokes, H. H. (2004) ‘On the advantage of using two or more econometric software systems to
solve the same problem’, Journal of Economic and Social Measurement 29: 307–320.
Swamy, P. A. V. B. and S. S. Arora (1972) ‘The exact finite sample properties of the estimators of
coefficients in the error components regression models’, Econometrica 40: 261–275.
Verbeek, M. (2004) A Guide to Modern Econometrics, New York: Wiley, second edn.
White, H. (1980) ‘A heteroskedasticity-consistent covariance matrix astimator and a direct test for
heteroskedasticity’, Econometrica 48: 817–838.
Windmeijer, F. (2005) ‘A finite sample correction for the variance of linear efficient two-step GMM
estimators’, Journal of Econometrics 126: 25–51.
Wooldridge, J. M. (2002a) Econometric Analysis of Cross Section and Panel Data, Cambridge, MA:
MIT Press.
Yalta, A. T. and A. Y. Yalta (2007) ‘GRETL 1.6.0 and its numerical accuracy’, Journal of Applied
Econometrics 22: 849–854.
Zhu, C., R. H. Byrd and J. Nocedal (1997) ‘Algorithm 778: L-BFGS-B: Fortran routines for large-scale
bound-constrained optimization’, ACM Transactions on Mathematical Software 23(4): 550–560.