Dear all,
Hope everyone is staying well in this difficult time.
I am bootstrapping standard errors of the estimates that are obtained from the following line of code:
npregress kernel price unitsize i.floor i.byear_lump, estimator(linear) kernel(gauss) dkernel(li)
Bootstrapping the standard error (using wild bootstrap) is straightforward, except that it takes a long time for each iteration to terminate.
I am running multiple instances of my code on a server, but the server imposes a time limit on how each instance can be run on it (24 hours), and it turned out that the code takes longer than that to terminate.
Other than reducing the number of observations (currently 72000, and the code terminates after 79000 seconds on my desktop), is there any way to get around the time limit, e.g., have Stata store the current state of the algorithm right before the time limit force quits the program, so that the Stata can restart from there when the code is run again?
Any advice would be greatly appreciated,
Thanks,
Sam
Related Posts with Getting around the time limit imposed by the server
Combining summary statistics and correlation matrixHi everyone, I need to create a table combining summary statistics and correlation matrix as a part …
How to deal with "factor variable and time-series operators not allowed "when I run gllamm command?Hello all, How can I include dummy variables in gllamm?Or how can I change factor variales into dumm…
Dealing with missing dataHi, I use the following equation for calculating return for a portfolio of stocks. gen MOMf3_p1= ( …
Using Kappa / Weighted KappaDear all Please advise if I can use kappa with this data structure: With two raters (fc and lp) sco…
Using Foreach with loop to calculate averageHi, I want to calculate average of a variable X for all households for each 200 categories on varia…
Subscribe to:
Post Comments (Atom)
0 Response to Getting around the time limit imposed by the server
Post a Comment