Hi,
I have two really large datasets (5,00,000 observations each) and I have been using the command "matchit id allNames using "xyz.dta", idusing(familyid) txtusing(allNamesFamily)" to match names between these two datasets.
However, matchit is taking a really really long time to carry out the fuzzy match (almost 24 hours). I have decided to run the same command but on smaller groups now however I am not sure how to create a loop function for it.
Essentially,
I want STATA to create district-level groups, carry out the above matchit command for each group, and save all the results together. The idea is that matchit only has to look for searches within each of these district-level sub-groups and not the whole dataset.
Can anyone help me with writing such a loop?
Related Posts with How to create a loop function to implement fuzzy matching on a large dataset?
-xtsfkk- and fixed effectDear Dr Karakaplan, thank you very much for providing the community with this new methodology. I'm…
-margins- following -mixed- commandHello, I would appreciate your help with estimation of -margins- following a -mixed- command. A sa…
Conditional logit (clogit) for cross-sectional dataDear all, I'm working on a project in which we analyze firms' location decision, i.e. the decision …
average at sector level within state (excluding current sector)Hi all, I have a company level data as below and I want to create a new variable that is the averag…
Error code in Google Places API--Observation numbers out of rangeI've been trying to run googleplaces to return latitude/longitude for some facilities I am analyzing…
Subscribe to:
Post Comments (Atom)
0 Response to How to create a loop function to implement fuzzy matching on a large dataset?
Post a Comment