An Analysis of Navy Recruiting Goal Allocation Models
An Analysis of Navy Recruiting Goal Allocation Models
CRM D0026005.A2/Final
December 2011
Approved for distribution: December 2011
This document represents the best opinion of CNA at the time of issue.
It does not necessarily represent the opinion of the Department of the Navy.
Introduction ....................................................................................... 5
Background and tasking .............................................................. 5
Issues ............................................................................................. 6
Approach ...................................................................................... 7
Organization of this report .......................................................... 8
ii
Ideas and suggestions ............................................................. 59
References .......................................................................................... 97
iii
This page intentionally left blank.
iv
Executive summary
Background and tasking
The Navy continually aims to have the right combination of per-
sonnel to meet its dynamic needs. Achieving the desired overall
force composition requires accessing the right mix of recruits—
enlisted and officer, active and reserve. To do this, the Navy must
have details about the available recruitable population, including
where specific types of people are located.
1
Approach
We began by talking to the winners of the Recruiter of the Year
(ROY) awards to understand what they think is lacking in the cur-
rent goaling process. Next, we reviewed the literature on recruiting
practices and modeling methods. Then, after reviewing how the
other services allocate their recruiting missions, we examined exist-
ing databases to see what data are available as model inputs. Finally,
we created a model for active enlisted personnel and identified what
data are necessary to create a similar model for the reserve enlisted
and active and reserve officers. Using information we collected
about officer data, we developed recommendations on how to im-
prove the current approach to officer recruit goaling.
Enlisted active
The current NRC model for goaling enlisted recruits for the active
component (AC) was rigorously developed. It uses much of the
publicly available data, but these data lack detail and specificity, par-
ticularly with regard to location. The goals are distributed to the
NRDs, which are fairly large. To provide NRC with the capability to
account for location in goaling, we develop a model for forecasting
the number of recruits of different types from each zip code. We
verify our model’s predictive capability and recommend adopting
the zip-code level model to make better use of available market
data. This more highly detailed goaling method will be helpful as
the NRC considers reorganizing and redistributing its goals both
geographically and demographically.
Enlisted reserve
As for reserve enlisted, NRC is using available data, but the data
have some of the same issues. For non-prior-service personnel, we
2
recommend using our active enlisted model with some minor
changes to data inputs.
Officer active
Officer Reserve
Medical
3
recruiters. One of our suggestions is to consider the “fantasy draft”
model for goaling medical recruiters, as is done in the Air Force.
4
Introduction
Background and tasking
Navy Recruiting Command (NRC) is responsible for recruiting four
main types of personnel, each of which constitutes a separate re-
cruiting market: officers and enlisted for the Active Component
(AC) and officers and enlisted for the Reserve Component (RC). In
addition, NRC has responsibility for recruiting all the Navy’s medi-
cal officers, an important submarket of AC and RC officers. An ef-
fective goaling process is a key factor in ensuring that all the Navy’s
accession goals are met and that recruiting resources are used effi-
ciently. In particular, the geographic allocation of the Navy’s re-
cruiting goals has a major impact on resource productivity, as well as
on the quantity, quality, and demographic mix of recruits. The goal-
ing processes for the five types of personnel listed above have not,
however, been updated or critically examined for some time. This is
significant because, over the last decade, several important changes
have occurred, including the following:
5
Issues
In general, the purpose of a goaling process is to distribute a total
service accession goal to recruiters in the field in a way that maxi-
mizes the probability that the aggregate mission is met. Depending
on its design, a goaling model may also be used to signal changes in
the recruiting environment that call for changes in the overall level
of recruiting resources or changes in their distribution across the
country.
6
tially valuable market information and the precise geographic allo-
cation of resources, including recruiters, advertising, and stations.
Approach
This study was done in two phases. We began with exploratory ac-
tivities, then, based on results from these activities, we moved to
model assessment and development.
Navy data
1
We summarize what we learned from ROY winners in appendix A.
7
— Personalized Recruiting for Immediate and Delayed
Enlistment (PRIDE)
Market data
8
Organization of this report
The paper is divided into five main sections, one for each market
segment (enlisted AC, enlisted RC, officer AC, and officer RC). The
final section looks at recruiting medical professionals. Within each
section, we describe the Navy’s current goaling method and identify
the issues we believe need to be addressed. This set up is followed by
a brief summary of the other Services’ goaling methods and lessons
learned from them. With all the necessary information in hand, we
then analyze the goaling issues and make recommendations.
9
This page intentionally left blank.
10
Enlisted active component (AC)
Introduction
The enlisted AC mission is by far the Navy’s largest recruiting mis-
sion and the goaling model for enlisted AC personnel is the Navy’s
most sophisticated. It uses a combination of statistical methods to
forecast high-quality male contract production. The model includes
historical production, recruiting resources, and economic and
population factors, as well as seasonality, pay, and other variables.
The Enlisted Goaling Model (EGM) is used to provide goals for
each of the two Regions (East and West). It also is also configured
to provide estimates of the recruiting potential for each of the 26
NRDs. The Regions may use the goaling model recommendations
for their NRDs, but they are free to modify these allocations. Each
NRD further redistributes its goals down to recruiting stations and
individual recruiters, using whatever approach they deem most rele-
vant.
The EGM has remained largely unchanged for at least two decades,
over which time the recruiting environment has changed substan-
tially. Two changes have particular relevance for enlisted AC goal-
ing. First, an increased emphasis on demographic diversity suggests
a need to include new factors in the model. Second, tighter budget
constraints, and the concomitant need to use the goaling model to
support efficient resource allocation, suggest a need to model em-
ploying a smaller unit of analysis.
11
from the goaling and modeling concerns of the other four types of
personnel addressed in this report.
Market definition
Geography
For the Navy, enlisted AC recruiting markets are defined
geographically by recruiting stations and the areas surrounding
them: together, the recruiters assigned to each station are
responsible for covering a surrounding geographic territory that is
roughly defined by zip codes [1]. The stations are nested within the
26 NRDs, and NRDs aggregate to the two Regions.
Eligibility
Within geographic areas, eligibility requirements for enlisted
personnel further define the enlisted AC market in terms of
education level and age. The primary target population for enlisted
AC recruiting is high school students and high schools graduates
ages 17 to 22. This is the primary market because members of this
age group are both “at the stage of life that career decisions are
natural” and at the “optimum training age” [1]. A secondary target
market consists of men in the 22- to 29-year-old age group, with or
without a high school diploma [1]. Although the secondary market
officially includes non-high-school-degree graduates (NHSDGs), the
Navy only enlists a limited number of these each year because this
group has been shown to have high first-term attrition relative to
2
high-school-degree graduates (HSDGs)[2].
2
The DOD restricts NHSDGs to 10 percent of total accessions, and the
Navy currently places a 5 percent cap on NHSDG accessions [30].
12
seek to enlist A- and Cu-cell recruits; the Navy does not enlist
anyone with an AFQT score below 35 and does not enlist NHSDGs
with AFQT scores below 50 [1].
Size
Between FY2007 and FY2011, the average size of the enlisted AC
3
recruiting mission was just under 36,000 accessions. With a mission
this large, it is feasible to use a statistical goaling model to allocate
the recruiting goals at a low level of geographic detail.
3
This data came from the Facts and Statistics tab on the CNRC website:
http://www.cnrc.navy.mil/PAO/facts_stats.htm .
4
To account for the possibility of a strong relationship between recruiting
in a particular quarter and recruiting in past quarters, the model is esti-
mated as an autoregressive form.
13
5
The number of production recruiters in each NRD
However, because the model was developed in the late 1980s and
early 1990s, the supply of “eligible recruits” refers to the supply of
male recruits only. Also, the model includes no information about
race and ethnicity.
Performance
The EGM was last evaluated in the late 1990s by CNA, which
reported its results in An Econometric Analysis of the Enlisted Goaling
Model by Goldhaber [4]. In general, Goldhaber found that the
model was reasonably accurate at predicting the number of A-cell
male enlistments on the NRD level, but that different models
5
The number of recruiters is determined using a different model. Each
year, planning staff at NRC determine the desired number of recruiters us-
ing a constrained optimization model that calculates the cost-minimizing
number of recruiters for a given beginning-of- year contract objective or
the maximum number of contracts for a given number of recruiters. The
model includes as parameters the programmed levels of other recruiting
resources (e.g., advertising and enlistment incentives), forecasts of the na-
tional unemployment rate and of military pay relative to civilian pay, as
well as the supply response to increases in the number of recruiters, which
comes from the EGM. [29]
14
should be used for subgroups of the population, such as workforce
recruits vs. high school seniors.
Unit of analysis
Based on our initial tasking and on additional conversations with
NRC staff, CNRC’s primary concern regarding the EGM is the unit
of analysis. Because NRDs are large—spanning multiple labor
markets and even multiple states—modeling at the NRD level does
not allow the Navy to precisely allocate the recruiting goal based on
market-specific conditions and needs. While recruiting goals are
eventually distributed down to the station level, this is done without
the help of a model or rigorous methodology. A goaling model with
15
more geographical detail could make both goal and resource
allocation more structured and potentially more efficient.
Demographic diversity
The Navy has become increasingly focused on demographic
diversity, with the goal of growing a force that is representative of
the nation in terms of race, ethnicity, and gender. And, compared
to the other Services, the Navy has been relatively successful at
recruiting racial/ethnic minorities and women into its enlisted
ranks. For example, the Military Leadership Diversity Commission
(MLDC) reported that, for 2007 and 2008, the Navy was the only
Service whose enlisted AC accessions were not disproportionately
white relative to the eligible recruiting pool. And, only the Air Force
had a higher share of women among its accessions for the same
years: 23 percent for Air Force accessions compared to just below 19
percent for Navy accessions. The female accession shares for the
Army and the Marine Corps were 16 and 7 percent, respectively. [5]
16
Lessons from other Services6
Before addressing these issues, we looked to the other Services for
ideas. Our review of the other Services’ goaling methods for the
enlisted AC market showed that none has a model that is as detailed or
statistically rigorous as the EGM. Instead, all three other Services
allocate their national goals to lower geographic levels using measures
of market size or past production, or combinations of the two.
The Army goals at the station level (there are 1,400 Army recruiting
stations across the country) use a weighted average of just two factors.
7
The first is a measure of historical past production that captures the
overall Department of Defense (DOD) production of high-quality con-
tracts. The second is a projection of the qualified military available
8
(QMA) population between 17 and 29 years of age. The weights the
Army assigns to these two factors can change over time. In the past, the
Army has assigned a 90 percent weight to past production and a 10
percent weight to the QMA population. Recently, however, the weights
changed to 60 percent and 40 percent, respectively. The Air Force al-
9
locates national recruiting goals at the group level based solely on a 5-
year average of past production, with the most recent three years
10
weighted more than the other two.
6
We are grateful to Mr. Mike Nelson at USAREC, Col T.J. Kenney at AFRS,
and Captain Joseph Wydeven at MCRC for providing CNA with the follow-
ing information on goaling in their respective services.
7
The past production measure is a weighted average of the last four years
of production, with the weights declining from 40 percent for the most re-
cent year to 30, 20, and 10 percent for the next three years.
8
QMA is the total 17- to 24-year-old population, excluding institutional-
ized and those in military service, unauthorized immigrants, and non-
HSDG not enrolled in high school or an equivalency program.
9
The Air Force Recruiting Service (AFRS) is organized into 3 groups, 24
squadrons, and 1,215 recruiting offices.
10
In the past, the model included manning and population data in addi-
tion to past production. In recent years, however, evaluations of the model
have shown the contribution of manning and population to be small in
comparison with past production; as a result, the Air Force has been using
the simplified model.
17
Finally, the Marine Corps goals at the zip-code level use two popu-
lation measures and a past production measure. It begins with es-
timates of the 17- to 24-year-old civilian non-institutionalized
population (CNIP) provided by an outside agency called Woods
and Poole (W&P). Then, it determines what percentage of enlist-
ees should score 50 or above on the AFQT using five years’ worth
of results from ASVAB (Armed Services Vocational Aptitude Bat-
tery) test takers (regardless of whether the test taker subsequently
enlisted). The population estimates are then combined with five
years’ worth of zip-code level DOD production data. Based on this
information, the Marine Corps determines what each station’s mis-
sion share should be. Furthermore, this population distribution is
used to determine the allocation of recruiting resources. If there is
a shift in population, recruiters are moved between the regions,
but the total remains the same.
18
Addressing the issues: A zip-code level goaling model with
diversity data
To address the issues raised by CNRC, we created a zip-code-level
model. Zip-code data are the most granular and detailed information
available about potential recruits. Including such data is beneficial
for several reasons. First, tracking specific population information by
zip code will allow the Navy to be more responsive to changing
demographic needs and to target specific subpopulations as neces-
sary. Like the Marine Corps, the Navy could begin allocating re-
sources and recruiters to those areas that are most likely to produce
particular types of contracts. Second, understanding which areas are
likely to be most and least productive could lead to the closing or
consolidation of recruiting stations. Similarly, the Navy could place
recruiting stations strategically to minimize the distance between
them and promising populations.
Research suggests that this final point on distance and travel time
between recruiting stations and potential recruits merits additional
attention. In 1992, Bohn and Schmitz looked at the effect that dis-
tance has on enlistment rates. They hypothesized that an increase in
distance would have a negative effect on the rate of recruitment and
found this to be true. [7] A later study by the same authors used a
different modeling technique but reached a similar conclusion: The
greater the distance from the recruiting station to a recruit’s loca-
tion, the more production rates fell. [8] Evaluating production on a
zip-code level can inform the NRC about where most of the recrui-
table population resides. This, in turn, can inform station place-
ment in an effort to decrease the distance and travel time between
the station and as many potential recruits as possible.
19
contracts at all, so we use a zero-inflated Poisson (ZIP) model. The
11
ZIP model provides a way of modeling the excess zeros in addition
to the counts of recruits we expect from each zip code.
Recent studies have shown the variables that go into our model as
predictors to be important for enlistments (see [10]). These vari-
ables are primarily socioeconomic in nature (e.g., population, edu-
cation, and crime data). Therefore, multicollinearity is a potential
12
modeling concern. Although this may not affect the model’s pre-
dictive ability, the estimated coefficients in the model can become
unstable and sensitive to model specification. The coefficients can
take signs that are counterintuitive, but, because the purpose of this
model is prediction rather than estimate coefficients, we consider
multicollinearity a secondary concern and caution against interpret-
ing the coefficients in isolation.
20
process over several years of data, the unemployment rate was not
shown to have a significant relationship with total enlistments, so we
did not include it in the model presented below, but it does come
up as significant in other models included in appendix B. Addition-
ally, we use market awareness measures as a substitute for propensity
to enlist and advertising.
IPEDS also provides the size of the nearest college or university. The
variable is divided into ordinal groups of under 1,000, 1,000–4,999,
5,000–9,999, 10,000–19,999, and 20,000 and over. We recoded these
as one though five, respectively, for our analysis. We hypothesize
that there might be a relationship between the size of the nearest
college or university and the academic atmosphere around the zip
code, and thus, the zip code’s likelihood of producing enlistees.
21
Historically Black College or University (HBCU)
Demographic data
22
Navy Awareness Index
CNRC provided us with the Navy Awareness Index (AI). This variable
measures each location’s awareness of the Navy by summarizing in-
formation on Navy leads. More specifically, AI estimates the likeli-
hood that consumers recognize the Navy’s “product” (i.e.,
employment/career opportunities available through naval service).
These data were measured by the Designated Market Area (DMA),
and the mapping between DMAs and zip codes was provided by
CNRC.
Recruiters
We include the number of recruiters from each service that are re-
cruiting in each zip code. Because a recruiter is often responsible
for more than one zip code, this number is sometimes a decimal.
These data are from RMIS.
Crime data
Veteran population
Model results
We ran our model for 2006 to 2010, with each model’s inputs pre-
dicting the following year’s recruiting results. Previous models have
been run using within-year data [10], but because our interest is
23
specifically in forecasting enlistments, we created a model that mir-
13
rors NRC’s work in this way.
Table 2 presents the results from the counts model. Of particular in-
terest are the variables with p-value less than 0.05, because those vari-
ables were deemed statistically significant, meaning that they correlate
well with our response variables, so for ease of reading, we have re-
moved those variables which were highly insignificant. Similar models
13
The Navy’s WEBSTEAM database, which is used by recruiters to identify
and analyze markets, could easily incorporate predictions from our model.
24
for subpopulations of the total inventory by race/ethnicity,
A-cell, and gender are presented in tables 7 through 13 in appendix B.
Table 2. Results from the Counts Model: Coefficients for modeling inventory on a zip-code
level
Variable Coefficient Standard Error p-value
Distance to responsible Navy Recruiting Station -0.0141 0.0005 0.0000
Distance to NRS squared 0.0000 0.0000 0.0000
W & P 17-19 currently in college Hispanic 0.0103 0.0057 0.0680
W & P 17-19 currently in college white -0.0152 0.0020 0.0000
W & P 17-19 currently in HS, year 1-3 black 0.0030 0.0012 0.0120
W & P 17-19 currently in HS, year 1-3 Hispanic 0.0141 0.0025 0.0000
W & P 17-19 currently in HS, year 1-3 white 0.0025 0.0013 0.0650
W & P 17-19 HSDG black 0.0011 0.0003 0.0000
W & P 17-19 HSDG Hispanic 0.0025 0.0004 0.0000
W & P 17-19 HSDG white 0.0003 0.0001 0.0040
W & P 17-19 senior in HS Hispanic -0.0113 0.0014 0.0000
W & P 17-19 senior in HS white 0.0017 0.0006 0.0040
W & P 20-21 college grad White 0.0075 0.0015 0.0000
W & P 20-21 currently in HS, year 1-3 black -0.0203 0.0060 0.0010
W & P 20-21 currently in HS, year 1-3 white -0.0292 0.0060 0.0000
W & P 20-21 HSDG black -0.0012 0.0003 0.0000
W & P 20-21 HSDG Hispanic -0.0015 0.0004 0.0000
W & P 20-21 HSDG white 0.0006 0.0001 0.0000
W & P 20-21 senior in HS black 0.0088 0.0023 0.0000
W & P 20-21 senior in HS Hispanic 0.0193 0.0029 0.0000
W & P 20-21 senior in HS white 0.0080 0.0016 0.0000
Navy Awareness Index in 2009 0.5996 0.0512 0.0000
USAF recruiters in 2009 0.0855 0.0182 0.0000
USMC recruiters 0.0325 0.0095 0.0010
USN recruiters 0.0677 0.0040 0.0000
Violent crime 0.0002 0.0001 0.0050
Property crime 0.0003 0.0000 0.0000
Veteran population 17-44 0.3459 0.0274 0.0000
Veteran population 65-84 -0.2061 0.0321 0.0000
Veteran population 85 and up 0.5167 0.0973 0.0000
Constant -1.6927 0.0637 0.0000
25
Distance to the nearest recruiting station, as well as its square, were
statistically significant in predicting total recruits in a zip code. As
expected, the distance to the nearest recruiting station has a nega-
tive relationship with enlistments. The farther the station is from a
given zip code, the less Navy presence is in that zip code, and the
more effort is required for the recruiter and the potential recruit to
make the contact. The square of the distance is also statistically sig-
nificant, but the coefficient is quite small, nearly zero, so it does not
carry practical importance because the coefficient has nearly no
impact on the predicted count of recruits.
26
The data on veterans is also significant. We note that the presence
of very young and very old veterans, who likely participated in the
Iraq War and World War II, respectively, has a positive relationship
with enlistments. The presence of those in the age group between
65 and 84 is negatively related, likely because of the controversial
war in Vietnam. The group of 45- to 64-year-olds is not significant,
possibly because both younger and older veterans are in that group.
On average, then, there is no effect. Unfortunately, these are the
only breakdowns provided by the dataset so we cannot investigate
this further.
Model diagnostics
Our next step was to compute the mean absolute deviation (MAD).
In the analysis described above, when we averaged the differences
between predicted and actual enlistments, negative and positive dif-
ferences canceled each other out, giving an optimistic estimate of
how accurate our model can be. We now know that it’s right on av-
erage, but we want to know by how many people we are wrong on
average. To do this, we first rounded the model predictions to the
14
Specifically, we used the Vuong test to check whether a regular Poisson
model could have been used rather than the zero-inflated one. The test gave
a p-value of 0, confirming that the zero-inflated model was appropriate.
27
15
nearest integer. We did so because this is how the NRC would use
the model: NRC would round the predictions to see how many
whole people it expects from each zip code. Then, we computed the
difference between actual enlistments and the predictions, but we
took the absolute value of the difference before averaging so that all
differences contribute to our MAD. We estimate the MAD to be
about 0.943. That is to say, our model is off by just less than one
person, on average, per zip code.
Recommendations
We developed this zip-code level model to help NRC distribute goal
in more geographical detail. After verifying the model’s predictive
capability in the previous section, we recommend adopting the zip-
code level model to make better use of available market data. The
improvement in the detail of this goaling method should prove
helpful as the NRC considers reorganizing, consolidating or closing
stations, and redistributing its goals both geographically and de-
mographically. Although the coefficients of the models should not
be interpreted in isolation, the models are valid for predictions, and
their forecasts can be used to determine the proportions and feasi-
bility of recruiting goals at each station.
15
In the zero-inflated Poisson model, the observed value is typically a
count, but a predicted value is a conditional mean (the average number of
events given the predictors) so it need not be an integer.
28
Specifically, we recommend the use of this detailed model in tan-
dem with the EGM, i.e. to use the EGM to come up with a total
NRD goal, and then use the zip-code model to allocate the EGM
goal to stations proportionately to the model predictions. On one
hand, this approach allows NRC to maintain their EGM which in-
cludes variables they deem important for enlistment, and which was
built on their substantial subject matter expertise. On the other
hand, the combination of the two models will allow a more detailed
adjustment for the recruiting markets and thus may improve the ef-
ficiency of resource use and equity among recruiter assignments.
29
This page intentionally left blank.
30
Enlisted reserve component (RC)
Introduction
There are two elements to RC recruiting: prior-service (PS) and
non-prior-service (NPS) recruiting. Recruiting for enlisted NPS re-
serves is similar to the AC because the target populations are the
same. In fact, recruiters can now recruit for both components, al-
though there are some major differences between the two. For in-
stance, because a much smaller number of people enlist in the
reserves, modeling and geographical allocation of resources are
more complex from a statistical perspective. In addition, because
reservists usually live at home and train on base at regular intervals,
recruiting for the RC depends on the location of drilling units, as
well as vacancies and authorizations— three restrictions that do not
come into play in modeling the AC. Because of the similarities in
target population for NPS recruits and AC recruits, the goaling
model is currently similar between the two, as are the concerns,
identified in the previous section.
31
Non-prior Service (NPS)
For NPS recruiting, NCR uses the New Accession Training goal
model, which is based on the same methodology as the goaling
model in the enlisted AC. (Please see the previous section for de-
tails.) The only difference is that populations that go into the reserve
model are constrained to residing within 50 miles of Navy Opera-
tional Support Centers (NOSCs), where the recruits undergo train-
ing. This is a logical restriction, given that recruits have to travel
regularly to train. The model predicts enlistments of A-cell men on
the NRD level and assumes that AC production recruiters recruit NPS
enlistees. The model has not been evaluated against production.
U.S. Army
The first step in the Army goaling process is to allocate all recruiters
geographically, regardless of their component (i.e., active, Army
16
We are grateful to Mr. Mike Nelson at USAREC, Colonel T.J. Kenney at
AFRS, and Captain Joseph Wydeven at MCRC for providing CNA with the
following information on goaling in their respective services.
32
National Guard, or Army reserve). This distribution is based on an
equal weighting of QMA and DOD past production. Subsequently,
recruiters are allocated by component. Active Guard Reserve (AGR)
recruiters are allocated separately from the rest of the recruiting
force, based on the following four weighted factors: 30 percent past
production, 20 percent vacancies, 40 percent TPU (troop program
or drilling unit) authorizations, and 10 percent QMA. The Army,
however, is currently considering a model that relies completely on
distributed authorizations and TPU structure as a fixed basis for lo-
cating AGR recruiters.
More precisely, the goal allocation model for recruiting NPS enlist-
ees is structured as follows:
25 percent authorizations
17
For current year yt, the weights are as follows: yt-1 = 40 percent yt-2 = 30
percent; yt-3 = 30 percent, and y1-4 = 10 percent.
33
currently over endstrength and require a precision mission.
Authorizations are weighted much lower than vacancies because at-
trition is currently very low. The weight given to past production is
low because past vacancy availability is likely to be different from fu-
ture vacancy availability. And, finally, because QMA is considered to
be a good measure of the NPS market, it remains in the model.
The PS enlisted model is built in the following way (the same units
are filtered out in the PS model as in the NPS model):
34
mobilization augmentee (IMA) billet locations in the continental
United States, Hawaii, and Puerto Rico. A 2001 CNA study by
Dolfini-Reed looks at how recruitable populations support the cur-
rent Selected Marine Corps Reserve (SMCR) force laydown [11].
The motivation for this work is that certain geographical areas have
difficulty supporting the manning requirements of some reserve
units. PS IMA billets have paygrade and skill requirements, particu-
larly because there is no training funding for IMA positions. This
makes recruiting for these billets more challenging than on the ac-
tive duty side, since location and skill matches limit who can be-
come a reservist. Furthermore, the travel policy states that a reservist
must live within 100 miles of the unit, or within three hours’ travel
time, whichever is less. In this sense, the USMC’s ability to man a re-
serve unit depends on its geographic location
35
Next, we consider the allocation of recruiters rather than reserve
units. CNA researchers Malone and Hattiangadi consider different
ways that the Marine Corps could allocate PS recruiters across dis-
tricts [12]. In the same document, CNA also offers two other meth-
ods for deciding where to place recruiters based on additional
variables. One of these variables is the “ease” of actually filling bil-
lets, which is defined as the number of leads divided by the number
of billets. A second variable that could be used in conjunction with
the first is the vacancy rate, which is defined as the number of leads
divided by the fill rate.
Recommendations
In considering the goaling methods laid out in this section, as well
as the enlisted model described in the previous section of this pa-
per, we provide some general recommendations for how the Navy
might improve its RC goaling process.
36
model, CNRC could address the issues discussed in the Marine
Corps section, such as merging and moving reserve units.
37
This page intentionally left blank.
38
Officer active component (AC)
Introduction
Compared with enlisted goaling, creating an officer goaling model
presents several additional challenges. First, the number of officer
accessions is smaller than the number of enlisted accessions, so sta-
tistical modeling is less reliable. Second, the geographic units of
analysis are more difficult to characterize. Typically, the services
consider colleges and universities their units of analysis rather than
zip codes or states. However, the Census and similar data sources,
which are readily available in fine geographic detail, may not apply,
since potential recruits may attend college far from home. When of-
ficers are recruited, their home of record is recorded inconsistently:
sometimes the services record the permanent residence and some-
times the college address. In addition, all-service officer data are not
as readily available as their enlisted counterpart. Instead, each ser-
vice has to use its own past production data.
39
we examine current methodology for officer recruiting, considering
the issues described above and discussing ways to improve goaling
methods. Given the noted data constraints, our analysis in this sec-
tion is necessarily qualitative—we rely on information gathered dur-
ing the Officer Goaling Conference, on discussions with recruiters
and recruiting leadership from the Navy and other services, and on
relevant literature. We also report on how the issue of officer diver-
sity in recruiting is approached in each service, and we make sug-
gestions for the Navy.
Description
The market measures consist of college degree data from the Na-
tional Center for Education Statistics (NCES). From this dataset, the
Navy focuses solely on Science, Technology, Engineering, or
Mathematics (STEM) degrees. The data are analyzed by the number
of degrees by degree type, race/ethnicity of the degree recipient,
gender of the degree recipient, and school quality. NRC excludes
nonresident aliens from these data and uses this as a proxy for citi-
zenship since no better citizenship data are available. CNRC maps
college majors to specific Navy officer designators, and goals can be
18
See:
http://www.cnrc.navy.mil/publications/Directives/1131%202E_CHA
TER%206_CH1.pdf
40
computed at this very detailed level. Overall, the Navy recruits into
67 broad officer categories, with a total of over 260 detailed goals.
Performance
Although NRC has not formally quantitatively evaluated the officer
goaling method, they are looking to improve their current model.
We first discuss the performance of the goaling method in general.
Then, we describe the problems associated with the Navy data. The
lack of data defines the limitations on our analysis.
Methodological issues
41
[6]. ROY awardees told us that, if a recruiter does an exceptional
job one year and recruits above and beyond her goal, she is tasked
with a higher number the following year. If she underrecruits, her
job next year will be easier. Past production is perceived as punish-
ing those who work harder and rewarding those who work less
19
hard . Second, in the case of officer recruiting, when the numbers
are small (e.g., one recruiter might recruit three or four people for
a given community), there are large chance variations in recruiting
outcomes. It is possible for someone to recruit four people for a
community one year, and nine people the next, without large
changes in the surrounding environment. The part of past produc-
tion that is due to chance alone can then affect future goals and set
up an expectation of eight people a year, which may not be a rea-
sonable expectation. Changes based on past production can be very
irregular and unsystematic. Third, relying on past production
means that, to some extent, the goaling expectations are always lag-
ging behind the changing recruiting environment.
Data issues
19
The caveat to this assessment is that, because the model smoothes over
the last five years of past production, the most recent year should not have
a strong effect on the change in goal. However, the model predicts on the
NRD level, and a district CO can use the most recent past production
numbers to alter a particular recruiter’s goal. Thus, although this large ef-
fect of recent production is not explicitly part of the goaling model, it is
part of the process, and is perceived negatively by recruiters.
42
which city produced the officer candidate. Finally, DOD-wide officer
accession data are not currently available. Obtaining DOD-wide
records will greatly increase the sample sizes for subsequent analysis,
making it more robust.
20
We are grateful to Mr. Mike Nelson at USAREC, Col T.J. Kenney at
AFRS, and Captain Joseph Wydeven at MCRC for providing CNA with the
following information on goaling in their respective services.
43
U.S. Army
The Air Force Recruiting Service (AFRS) recruits general duty offi-
cers who do not come from ROTC or the Air Force Academy. These
usually represent between 10 and 20 percent of the general duty of-
ficers who go to Officer Training School (OTS). Senior enlisted
personnel with past recruiting experience recruit all officers in the
Air Force, with the exception of chaplains and lawyers, who are re-
cruited by volunteer recruiters in their ratings.
The Air Force officer recruiting goals are assigned based on pro-
pensity to enlist (as measured by past production) and manning.
This suggests that market factors are taken into account, albeit indi-
rectly. The Air Force places importance on recruit quality: all can-
didates are required to take the Air Force Officer Qualifying Test
44
(AFOQT). This test, like the Graduate Record Examination (GRE),
covers academics, analytical thinking, and mathematics. It contains
five composites: pilot, navigator, academic, verbal, and quantitative.
The test measures aptitudes, and scores are later used to select can-
didates for commissioning programs, such as OTS or Air Force Re-
serve Officer Training Corps (AFROTC). It is also used for selection
into specific training programs, such as pilot and navigator training.
In addition to the AFOQT, an officer applicant needs at least a 3.0
grade-point average (GPA) to be competitive.
45
The Marine Corps uses the QCP report to determine where Officer
Selection Officers (OSOs) should focus their attention for recruit-
ing on college campuses [15]. Given that there are only about 70
OSOs at 60 or so Officer Selection Stations (OSSs), the Marine
Corps must identify where the QCP is concentrated, and assign
OSOs accordingly, in order to optimize the use of their recruiting
resources. The market of potential Marine Corps officer candidates
includes people who are test-score qualified and either have a
bachelor’s degree or are in the process of earning one. This makes
schools a good area to focus on when deciding where to assign
recruiters.
Finally, the QCP paper [14] also presents an analysis that augments
the traditional QCP data with propensity to enlist and medical data
collected by the Centers for Disease Control and Prevention (CDC).
This analysis results in more precise estimates of the QCP popula-
tion, and generates a refined list of the schools with the most officer
candidates. This kind of additional market information could help
the Navy better evaluate its officer recruiting market.
Although officer racial and ethnic diversity is a concern for the Ma-
rine Corps, recruiting at Historically Black Colleges and Universities
(HBCUs) hasn’t been very successful. In addition, these colleges do
not usually show up high on the ranked QCP list. Rather, there is a
46
separate list of QCP schools for minority officers [14]. In the Ma-
rine Corps, each OSO must submit a certain number of diversity
applications based on feasibility, as computed by QCP estimates.
Anecdotally, however, because only applications are currently
goaled, recruiters tend to submit candidates that they know will not
get in simply to accomplish that goal.
21
The Navy’s minimum SAT score requirement is 1050, whereas it is 1000
for the USMC, so the list is slightly different for the Navy. In addition, we
did not use Barron’s data to identify how competitive the schools are; we
used the default measurement provided by IPEDS.
47
and medical data. To construct this table, we used data from easily
accessible online sources. We used the IPEDS dataset to obtain the
list of schools, their competitive attributes, and enrollment data.
48
ity enrollment. The fourth column is the total number of full-time
undergraduate students, and the two columns that follow are the
numbers of black and Hispanic students. These numbers are ob-
tained by multiplying the respective proportions in columns 2 and 3
by the total enrollment number. Table 4 is the QCP table created
for the Marine Corps, with adjustments made for medical data and
propensity to enlist. There are many similarities between the quick
data pull and the original QCP school list, indicating that this is an
inexpensive way to get at roughly correct data. Note that this list
does not focus on any specific kind of degree, but rather on the to-
tal number of students. As we will see later, this method is less useful
when only STEM degrees are of interest.
When we looked at the officer data for the Navy and compared the
non-missing school entries with top QCP schools that we identified
earlier, there was little correlation between what schools came out
as top QCP schools and where most Navy officers came from. This is
likely because the QCP analysis that CNA performs for the Marine
49
Corps does not make adjustments for STEM degrees. We present
our comparison in table 5. The third column indicates rank based
on total QCP population. The rightmost column indicates where
the QCP schools rank when sorted by Navy inventory.
Full-time
undergraduate IPEDS Navy
Institution name enrollment QCP rank rank
Arizona State University 45,597 1 27
Ohio State University 37,864 2 27
Pennsylvania State University 37,485 3 15
Texas A&M University 35,400 4 6
The University of Texas at Austin 35,364 5 3
University of Central Florida 34,197 6 541
Michigan State University 33,429 7 32
University of Florida 31,316 8 4
Indiana University–Bloomington 31,061 9 36
University of Illinois at 30,639 10 19
Urbana–Champaign
Purdue University–Main Campus 30,334 11 60
University of Minnesota 28,539 12 27
University of Washington–Seattle 28,094 13 16
Brigham Young University 28,048 14 24
Florida State University 27,705 15 11
Rutgers University–New Brunswick 27,588 16 114
University of Wisconsin–Madison 27,386 17 12
University of Arizona 27,103 18 67
University of California, Los Angeles 25,772 19 102
University of Michigan–Ann Arbor 25,342 20 13
University of Colorado at Boulder 24,916 21 24
University of California–Berkeley 24,797 22 2
University of Georgia 24,669 23 38
University of Maryland 24,617 24 5
Temple University 24,114 25 36
50
and those colleges and universities where the Navy has successfully
recruited in the past. Thus, the Marine Corps approach is not likely
to work for Navy goaling.
51
These additional market characteristics could help isolate the most
likely candidates for the Navy from within college populations as
well.
As was done in the recent CNA study for the Marine Corps
[14], we suggest that Joint Advertising Marketing Research &
Studies (JAMRS) estimates of youth propensity to enlist be in-
cluded in the NRC model. As we showed in table 4, in the case
of USMC, including this information changed the school order
on the ranked list. Including propensity may also substitute for
the current model’s reliance on past production.
52
Also based on the Malone et al. study, we suggest that NRC es-
timates be adjusted for data from the CDC [14]. The data
available online include diabetes, obesity, and activity infor-
mation on a county level. Including this information will have
an effect on the order of the ranked schools.
Diversity
53
applicants. The recruiting region could be recognized for
achieving that balance. If the minority selection rate is higher
than that for nonminorities, the NRD could be recognized for
achieving outstanding quality in minority applications. Both
this suggestion and the previous one would fare well with the
Navy’s recent proposal to goal applications for all officers.
54
Officer reserve component (RC)
Introduction
As on the enlisted side, there are two types of recruits who enter the
officer reserves: those with prior service (PS) and those without
(NPS). Those with prior service enter the reserves with a certain
skill set and training and are usually looking for a particular job.
They are usually recruited within a short amount of time after com-
pleting an active duty obligation. In contrast, NPS officers who
don’t come in through a Navy program such as ROTC, enter the re-
serves in a similar way as the active duty in the sense that they are
recruited out of the general college-educated population and with-
out prior Navy training. The recruiting process is quite different for
NPS and PS officers, so we address the two groups separately in this
section.
Officer reserve PS
Current goaling model
One issue we identified with the RC officer goal model is that its
heavy reliance on past production makes recruiters distrust it. How-
ever, one minor benefit of using the last five years of production is
55
that one year of increased production has a relatively small impact
on the predicted mission. Hence, this system should not be as de-
motivating as one with a heavier reliance on most recent numbers.
Nevertheless, the five-year window for evaluating production could
lead to a risk that CNRC can lag behind market changes by several
years.
The IRR list, which is given to recruiters, has records for those who
left the military in the last three years. It is not clear, however, if this
is the optimal time span. Although we could not locate existing
studies on how long someone should stay on that list, at the Officer
Goaling Conference it was discussed that three years is too long and
that, after a year or two, there is little chance of someone from the
IRR list joining the reserves.
22
We are grateful to Mr. Mike Nelson at USAREC, Col T.J. Kenney at
AFRS, and Captain Joseph Wydeven at MCRC for providing CNA with the
following information on goaling in their respective services.
56
areas in accordance with the allotted goal. Like the Navy, the Air
Force keeps the Individual Ready Reserve (IRR) list. The IRR list
gets populated during the exit interview that each active duty officer
must have with a recruiter, and the Air Force sends out quarterly
mailings to that address. The Air Force experiences problems with
the addresses on the IRR list; they are frequently inaccurate because
officers leaving the Air Force either don’t go to the location they
planned or they move.
To help mitigate similar issues, the Army has a special division re-
sponsible for management of the IRR list, under the Chief of Army
Reserves.
57
presented in the section on enlisted recruiting) which model
predicts the outcome the best. There will be inevitable en-
dogeneity in this type of a test since, intuitively, the recruiting
goal can affect the recruiter’s behavior. However, recruiters
have told us that, while they take the goal into account, it is not
the main driver of their work, so there might be some validity
to a test like this.
58
The Army and the Air Force recruit very few NPS officers into the
RC so their recruiting and goaling methods are not systematic. In
fact, the Air Force does not goal PS and NPS separately because it
has a strong preference for PS officers. Those interested in becom-
ing NPS officers in the Air Force have to interview for the position,
and typically there are more leads than billets, so the Air Force does
not aggressively recruit NPS officers.
59
This page intentionally left blank.
60
Medical officer recruiting
Background
Across services, medical recruiting is the most difficult type of re-
cruiting. Doctors frequently have to give up a convenient schedule,
a large salary, and sometimes a private practice to join the military.
They also spend more time in school in pursuit of their degrees
than most other recruits.
The study’s authors posit that increasing accessions can only hap-
pen as the result of services providing additional resources to en-
hance and expand current programs, as well as to research the
practices of other services [17].
61
growing, which may also have a negative effect on recruiting medi-
cal personnel into the military [18].
U.S. Navy
MC students
DC students
MSC students
NC students
62
and historical production over the last five years is weighted at 40
percent.
The workforce data are from PUMS, specifically the U.S. Census
ACS, which gives the geographic locations, by state, for the medical
workforce age 18 to 40. The main potential issue with these data is
that they are organized on a state level, which is fairly large and
does not always correspond with the NRD assignments. Still, this is
the only breakdown available in the ACS. The other two compo-
nents of the model are recruiter and historical five-year production,
each weighted at one-third.
When recruiting DCOs for the RC, the Navy usually looks for
graduate students, as well as individuals in residency programs, Cer-
tified Registered Nurse Anesthetist (CRNA) programs, and trade
organizations. The goal is 20 percent market (as measured by the
medical workforce), 40 percent designated reserve recruiters, and
63
40 percent historical production. NAVETs are usually recruited di-
rectly from active duty or from the IRR lists.
Perhaps the key issue with respect to setting goals is how goals affect
the motivation and productivity of recruiters. We heard from several
recruiters and from leadership that competition, incentives, and re-
wards might be the best way to allocate Navy medical recruiting
goals. They believe that the interest and motivation generated by a
competitive environment will overshadow the unfairness that comes
from assigning goals in small numbers.
23
We are grateful to Mr. Mike Nelson at USAREC, Colonel T.J. Kenney at
AFRS, and Captain Joseph Wydeven at MCRC for providing CNA with the
following information on goaling in their respective services.
64
graduating classes from medical school would be a useful addition
to the way the Navy currently goals medical officers.
The Air Force, like the Army, relies on senior enlisted recruiters to
recruit medical personnel. For medical goals, the Air Force uses
both applications and board success. However, they monitor board
results to make sure quality of applicants is competitive. The AFRS is
the major recruiter for medical professions. It recruits 90 percent of
all health care professionals (medical doctors, nurses, Medical Ser-
vice Corps, etc.). One of the factors considered in medical goals is
feasibility of recruiting numbers given recruiting resources. Goals
come from officer endstrength projections of requirements, which
are determined independently from recruiting resources. Hence,
goals may not be feasible. For example, two years ago the AFRS was
tasked with recruiting a large number of fully qualified MDs, but
had resources allowing for only a few dozen medical recruiters. This
year they had a mission of 25 fully qualified MDs, which was achiev-
able. Next year it will be 55, which is also reasonable. The Air
Force’s workforce medical doctor goals are by specialty, as are the
Navy’s.
Until recently, in the Air Force, the allocation of medical goal was
done based on past production, without much emphasis on particu-
lar medical subspecialties. For example, it recruited surgeons, in
general, as opposed to surgeons with particular specializations.
Starting in FY 2011, however, the requirements became more de-
tailed. To distribute these newly specific goals, for FY 2012, the allo-
cation of many of the more challenging medical goals (fully
qualified MDs) will happen using a “fantasy draft” model. Each
group gets to select one of the goals until all the goals have been al-
located. The group gets the entire nation to recruit against these
goals. For example, if a recruiter is looking for an oral surgeon, he
or she can go to conventions and visit places across the country to
find the appropriate specialty. If a different group finds an oral sur-
geon, it has to refer that person to the group tasked with recruiting
that specialty.
65
Potential ways forward
In reviewing the Navy process and considering the other services, we
identified two problem areas in the current goaling process: one is
market related, and the other is process related. Thus, we present
two sets of considerations for a way forward. The first assumes that
the current goaling process will continue to be used, but aims to
improve it by adding additional market data. The second explores
the idea of completely restructuring the medical goaling process by
introducing competition and incentives as tools to motivate recruit-
ers. In appendix D, we present an overview of existing literature on
recruiter incentives, which can apply to both sets of considerations.
66
that order, each participant would choose a medical job category to
recruit and would be solely responsible for recruiting into that cate-
gory regardless of the region of the country. A competition could be
introduced based on completing the recruiting task. Alternatively,
the medical professions could be ranked based on difficulty to re-
cruit, and an auction draft could be implemented in which the par-
ticipants could trade their medical recruiting missions. NRC could
wait to see whether this model works successfully for the Air Force
or, alternately, conduct a study to assess this approach’s potential
for success.
67
This page intentionally left blank.
68
Future work
In this section, we briefly describe three potential extensions of the
analysis developed in this study.
69
uncertainty prevents improvements to the current goaling
model because it limits our knowledge of where the officers
are actually being recruited.
70
Appendix A: Recent goal allocation concerns
and ROY winner interviews
Every year, the Navy presents 13 to 15 top enlisted and officer re-
cruiters with the Recruiter of the Year, or ROY, award. At the be-
ginning of the project, we had an opportunity to talk with some of
last year’s winners about the current recruiting environment. Those
discussions helped inform this project. Although it wasn’t feasible to
study all topics identified during those discussions, we describe the
most salient here for reference.
Diversity recruiting
Recruiters who won awards for their diversity numbers mentioned
that they are more successful in inner-city areas, and that it helps if
the recruiter is of the same gender and racial/ethnic background as
the potential recruits. They mentioned having to travel quite a bit to
achieve their numbers, and they talked about the use of social me-
dia tools to attract recruits. They also said they don’t specifically re-
cruit to the recommended diversity targets: they do their job every
day and diversity takes care of itself. This emphasizes their reliance
on CNRC’s ability to geographically allocate diversity goals. Recruit-
ers mentioned a recent emphasis on quality rather than quantity
and a push for women, Hispanic, and African American A-cell
recruits.
Merging missions
With more extensive use of the reserves in wars in Iraq and Afghani-
stan, recruiting into the reserve component (RC) has become more
critical. Also, in a cost-constrained environment, the Navy is merg-
ing missions in its four major markets (i.e., enlisted, officer, AC, and
RC), enabling reassignment of recruiters across missions, and, in
71
some instances, a recruiter to recruit for all four. This could intro-
duce cost savings by closing some of the recruiting stations.
In addition, the Navy has slowly been merging enlisted and officer
missions. In our interviews, recruiters did not express concerns
about merging active and reserve missions; in fact, they explained
that this is already being done to some degree. However, merging
enlisted and officer missions seemed to generate more concern. On
the positive side, the recruiters thought that merging these missions
would enable officer candidates to come into any office and apply
without having to go to officer recruiting. Because parts of the re-
cruitable population are similar for enlisted and officer missions,
recruiters would also be able to cross-reference high schools and
colleges. In fact, the recruiters told us that many candidates are not
sure which of the two careers they prefer when they first consider
joining the Navy.
72
enlisted application takes only a few days. This will make the proc-
esses difficult to merge.
Recruiters currently use STEAM to learn about their market and de-
termine how to allocate their time. For additional information, they
reach out to recruiters who have worked in the same area, and they
investigate their assigned neighborhoods for socioeconomic factors,
such as income and political affiliation.
Technology
The technology of recruiting has changed markedly over the last 20
years, with new marketing technologies, social networking, and
greater recruiter mobility. Of the many recent technological ad-
vances, those making it easier to keep in touch have had the most
impact on Navy recruiting.
73
Recruiters discussed several tools that have enhanced their produc-
tivity, as well as technologies they see as necessary for improving
their workflow. Several recruiters mentioned that they use Facebook
to communicate with potential recruits and those in the Delayed
Entry Program (DEP). Many find it easier to contact recruits via
Facebook than by phone, and they use the “Friends of Friends” fea-
ture in Facebook to find additional candidates and get references.
This technique allows them to indentify candidates who are mem-
bers of swim or other athletic teams. Some also use Facebook to
check that potential recruits adhere to the Navy’s height and weight
standards, and they use this information to determine qualified
candidates or to provide mentoring. Recruiters are also finding that
texting rather than calling young recruits results in more frequent
interaction.
Because social networking sites and text messaging have become the
leading ways that young people communicate, inability to text or go
on Facebook hurts recruiters. Regular contact with those in the De-
layed Entry Program and mentees is important, and both social
networking and texting help recruiters stay in touch. Most recruiters
expressed dissatisfaction with the amount of paperwork and with
the number of computer- and phone-related issues they encounter.
Because of frequent travel, they mentioned the need for portable
computers and printers.
Past production
Several recruiters told us that they rely on CNRC to give them
achievable goals, and they direct their time toward recruiting in
general rather than toward the specific number provided by
24
CNRC. Currently, past production (i.e., the number of recruits
brought on in a district in the previous several years) makes up a
portion of the model that forecasts enlistments for each area. This is
not surprising because past production is a good measure of the re-
cruiters’ market, and recruiters agree that without past production
informing some of the goaling model, the missions would become
stagnant.
24
CNRC actually only goals Regions. NRD COs then goal recruiters.
74
The problem with this measure, however, is that it does not neces-
sarily incentivize recruiters in the right way. One’s hard work one
year may result in a harder task the following year, whereas poor
production could be “rewarded” with a smaller, more accomplish-
able goal. One recruiter told us that he was goaled one year with
four officers in a particular community, but delivered nine, so the
following year, he was goaled with nine, which was quite hard to ac-
complish. Recruiting models use more than one year of past pro-
duction (typically three to five), so this sharp increase in required
numbers is likely a result of the goaling process (leadership over-
sight) rather than of the goaling model alone. Nevertheless, re-
cruiters suggested that goals should be raised in a marginal manner
and based on the nation’s need, rather than on past production.
75
This page intentionally left blank.
76
Appendix B: Enlisted active component (AC)
zip-code-level model results
This appendix presents resulting model estimates for modeling re-
cruit production on a zip-code level for various subpopulations of
the total inventory. Several models were unstable with respect to
model specifications, and some were impossible to estimate. We
present, for each subpopulation, the best estimable model.
A-cell recruits
Table 7 presents our model results for predicting A-cell recruits by
zip code.
77
W & P 20-21 currently in HS, year 1-3 white -0.0298 0.0067 0.0000
W & P 20-21 HSDG black -0.0010 0.0003 0.0030
W & P 20-21 HSDG Hispanic -0.0014 0.0004 0.0010
W & P 20-21 HSDG white 0.0005 0.0001 0.0000
W & P 20-21 senior in HS, black 0.0041 0.0026 0.1120
W & P 20-21 senior in HS, Hispanic 0.0243 0.0033 0.0000
W & P 20-21 senior in HS, white 0.0073 0.0018 0.0000
Navy Awareness Index in 2009 0.6900 0.0571 0.0000
USAF recruiters in 2009 0.0743 0.0204 0.0000
USA recruiters 0.0071 0.0033 0.0350
USMC recruiters 0.0311 0.0105 0.0030
USN recruiters 0.0699 0.0044 0.0000
Violent crime 0.0002 0.0001 0.0090
Property crime 0.0002 0.0000 0.0000
Veteran population 17-44 0.3288 0.0306 0.0000
Veteran population 45-64 0.0091 0.0314 0.7720
Veteran population 65-84 -0.1408 0.0336 0.0000
Veteran population 85 and up 0.4308 0.1027 0.0000
Constant -1.9588 0.0706 0.0000
------------- ----------- ----------- --------
Non-zero model
Distance to closest college / university 0.1724 0.0135 0.0000
Distance squared -0.0023 0.0002 0.0000
College or university in zip -0.2252 0.2142 0.2930
Size of the college / university -0.1578 0.0478 0.0010
Size by distance interaction -0.0008 0.0025 0.7390
Multiple school flag -2.7610 2.7492 0.3150
Historically black college or university in zip -0.0158 0.1943 0.9350
Constant -1.9235 0.2084 0.0000
Black recruits
This model was unstable and not robust to model specifications. In
many specifications, our software was unable to estimate the model.
We present one of the successful specifications in table 8. For brev-
ity, we omit the statistically insignificant results.
78
Table 8. Model results for predicting the number of black recruits by zip code
Variable Coefficient Standard Error p-value
Distance to NRS -0.0118 0.0018 0.0000
State unemployment rate 0.0296 0.0117 0.0120
QMA black 17-19 0.0078 0.0021 0.0000
QMA white 17-19 -0.0079 0.0012 0.0000
W & P 20 no high school black -0.0039 0.0014 0.0060
W & P 20 GED black 0.0671 0.0238 0.0050
W & P 20 GED white 0.0319 0.0107 0.0030
W & P 20 in HS, years 1-3 black 0.0869 0.0313 0.0060
W & P 20 in HS, years 1-3 white 0.0786 0.0401 0.0500
W & P 20 HSDG -0.0102 0.0024 0.0000
W & P 20 Native American no HS 0.0041 0.0014 0.0030
W & P 20 HS senior black -0.0263 0.0134 0.0490
W & P 20 HS senior white 0.0645 0.0145 0.0000
W & P 21 college grad black -0.0929 0.0310 0.0030
W & P 21 no HS black 0.0048 0.0013 0.0000
W & P 21 GED white -0.0326 0.0126 0.0100
W & P 21 in HS years 1-3 black -0.0822 0.0297 0.0060
W & P 21 HSDG Hispanic -0.0100 0.0043 0.0200
W & P 21 no HS black -0.0056 0.0026 0.0340
W & P 21 AA degree white 0.0194 0.0096 0.0430
W & P 21 senior in HS black 0.0385 0.0138 0.0050
W & P 21 senior in HS Hispanic 0.0696 0.0230 0.0030
W & P 21 senior in HS white -0.0466 0.0133 0.0000
W & P 22 college grad black -0.0066 0.0026 0.0120
W & P 22 HSDG black 0.0056 0.0023 0.0140
W & P 22 HSDG Hispanic -0.0096 0.0052 0.0630
W & P 22 HSDG white 0.0052 0.0021 0.0150
W & P 17-19 GED black -0.0232 0.0107 0.0300
W & P 17-19 GED Hispanic -0.0458 0.0261 0.0800
W & P 17-19 HS years 1-3 black 0.0139 0.0084 0.0980
W & P 17-19 years 1-3 Hispanic 0.0028 0.0010 0.0050
W & P 17-19 years 1-3 white 0.0077 0.0019 0.0000
W & P 17-19 in college white 0.0018 0.0005 0.0010
W & P 17-19 senior in HS Hispanic -0.0103 0.0047 0.0300
W & P 23-24 college grad black 0.0062 0.0012 0.0000
W & P 23-24 AA degree black -0.0056 0.0030 0.0620
W & P 23-24 AA degree Hispanic 0.0094 0.0055 0.0900
W & P 23-24 senior in HS black 0.0541 0.0194 0.0050
W & P 23-24 senior in HS Hispanic -0.0999 0.0279 0.0000
79
Navy Awareness index 0.8029 0.1613 0.0000
USAF recruiters 0.1965 0.0472 0.0000
USA recruiters -0.0140 0.0084 0.0950
USMC recruiters 0.0872 0.0293 0.0030
USN recruiters 0.0135 0.0112 0.2300
Violent crime -0.0005 0.0002 0.0070
Property crime 0.0004 0.0000 0.0000
Veteran population 45-64 0.5852 0.0857 0.0000
Veteran population 65-84 -1.0567 0.1151 0.0000
Veteran population 85 and up 1.1917 0.3386 0.0000
Constant -3.6337 0.2289 0.0000
------------- ----------- ----------- --------
Non-zero model
College or university in zip -3.2560 0.4215 0.0000
Constant 1.5061 0.5197 0.0040
Table 9. Model results for predicting the number of black A-cell recruits by zip code
Variable Coefficient Standard Error p-value
Distance to NRS -0.0183 0.0023 0.0000
Distance to NRS squared 0.0000 0.0000 0.0000
W & P 17-19 currently in college black 0.0300 0.0142 0.0350
W & P 17-19 currently in college Hispanic -0.0556 0.0213 0.0090
W & P 17-19 currently in college white -0.0285 0.0089 0.0010
W & P 17-19 currently in HS, year 1-3 black -0.0049 0.0026 0.0570
W & P 17-19 currently in HS, year 1-3 Hispanic 0.0152 0.0096 0.1130
W & P 17-19 currently in HS, year 1-3 white 0.0114 0.0050 0.0220
W & P 17-19 HSDG black 0.0006 0.0006 0.2830
W & P 17-19 HSDG Hispanic 0.0027 0.0015 0.0670
W & P 17-19 HSDG white -0.0013 0.0004 0.0040
W & P 17-19 senior in HS, black 0.0072 0.0016 0.0000
W & P 17-19 senior in HS, Hispanic -0.0161 0.0052 0.0020
W & P 17-19 senior in HS, white -0.0035 0.0023 0.1250
W & P 20-21 college grad black 0.0056 0.0140 0.6880
W & P 20-21 college grad Hispanic 0.0284 0.0089 0.0010
80
W & P 20-21 college grad white 0.0031 0.0060 0.5990
W & P 20-21 currently in HS, year 1-3 black -0.0596 0.0187 0.0010
W & P 20-21 currently in HS, year 1-3 Hispanic -0.0076 0.0179 0.6710
W & P 20-21 currently in HS, year 1-3 white 0.0081 0.0218 0.7090
W & P 20-21 HSDG black -0.0008 0.0007 0.2030
W & P 20-21 HSDG Hispanic -0.0017 0.0015 0.2500
W & P 20-21 HSDG white 0.0020 0.0004 0.0000
W & P 20-21 senior in HS, black 0.0231 0.0068 0.0010
W & P 20-21 senior in HS, Hispanic 0.0318 0.0100 0.0010
W & P 20-21 senior in HS, white 0.0052 0.0060 0.3840
Navy Awareness Index in 2009 1.4917 0.1763 0.0000
USAF recruiters in 2009 0.2188 0.0540 0.0000
USA recruiters 0.0080 0.0091 0.3830
USMC recruiters 0.0217 0.0311 0.4860
USN recruiters 0.0268 0.0124 0.0300
Violent crime 0.0002 0.0002 0.2020
Property crime 0.0004 0.0000 0.0000
Veteran population 17-44 0.3167 0.0866 0.0000
Veteran population 45-64 0.4756 0.0967 0.0000
Veteran population 65-84 -1.0550 0.1266 0.0000
Veteran population 85 and up 1.0796 0.3744 0.0040
Constant -4.9429 0.2238 0.0000
------------- ----------- ----------- --------
Non-zero model
Distance to closest college / university 0.1712 0.0309 0.0000
Distance squared -0.0016 0.0005 0.0010
College or university in zip -0.1029 0.2970 0.7290
Size of the college / university -0.0248 0.0786 0.7520
Size by distance interaction -0.0038 0.0058 0.5120
Multiple school flag -1.9223 1.9512 0.3250
Historically black college or university in zip -2.2667 0.7687 0.0030
Constant -1.4263 0.3712 0.0000
Hispanic recruits
Table 10 presents our model results for predicting the number of
Hispanic recruits by zip code.
81
Table 10. Model results for predicting the number of Hispanic recruits by zip code
Variable Coefficient Standard Error p-value
Distance to NRS -0.0313 0.0015 0.0000
Distance to NRS squared 0.0000 0.0000 0.0000
QMA 17-19 I-IIIA black 0.0022 0.0003 0.0000
QMA 17-19 I-IIIA Hispanic 0.0056 0.0002 0.0000
QMA 17-19 I-IIIA white 0.0006 0.0000 0.0000
Navy Awareness Index in 2009 1.7594 0.1071 0.0000
USAF recruiters in 2009 0.1194 0.0362 0.0010
USA recruiters 0.0317 0.0052 0.0000
USMC recruiters 0.1885 0.0162 0.0000
USN recruiters 0.0764 0.0072 0.0000
Violent crime 0.0010 0.0001 0.0000
Property crime 0.0003 0.0000 0.0000
Veteran population 17-44 0.4855 0.0674 0.0000
Veteran population 45-64 -0.3404 0.0712 0.0000
Veteran population 65-84 -0.1335 0.0770 0.0830
Veteran population 85 and up 0.9779 0.2252 0.0000
Constant -4.0439 0.1407 0.0000
------------- ----------- ----------- --------
Non-zero model
Distance to closest college / university 0.2194 0.0189 0.0000
Distance squared -0.0036 0.0004 0.0000
College or university in zip 0.4802 0.1201 0.0000
Size of the college / university -0.0059 0.0353 0.8670
Size by distance interaction -0.0106 0.0031 0.0010
Multiple school flag -0.1532 0.1733 0.3770
Historically black college or university in zip 0.7875 0.1904 0.0000
Constant -1.1273 0.1785 0.0000
82
Table 11. Model results for predicting the number of Hispanic A-cell recruits by zip code
Variable Coefficient Standard Error p-value
Distance to NRS -0.0338 0.0017 0.0000
Distance to NRS squared 0.0000 0.0000 0.0000
QMA 17-19 I-IIIA black 0.0021 0.0003 0.0000
QMA 17-19 I-IIIA Hispanic 0.0054 0.0002 0.0000
QMA 17-19 I-IIIA white 0.0007 0.0000 0.0000
Navy Awareness Index in 2009 1.8643 0.1187 0.0000
USAF recruiters in 2009 0.1421 0.0402 0.0000
USA recruiters 0.0366 0.0059 0.0000
USMC recruiters 0.1959 0.0180 0.0000
USN recruiters 0.0747 0.0079 0.0000
Violent crime 0.0009 0.0001 0.0000
Property crime 0.0003 0.0000 0.0000
Veteran population 17-44 0.4212 0.0738 0.0000
Veteran population 45-64 -0.2957 0.0771 0.0000
Veteran population 65-84 -0.1115 0.0809 0.1680
Veteran population 85 and up 0.9464 0.2329 0.0000
Constant -4.3508 0.1543 0.0000
------------- ----------- ----------- --------
Non-zero model
Distance to closest college / university 0.2254 0.0213 0.0000
Distance squared -0.0036 0.0004 0.0000
College or university in zip 0.4886 0.1349 0.0000
Size of the college / university 0.0324 0.0394 0.4100
Size by distance interaction -0.0123 0.0036 0.0010
Multiple school flag -0.0299 0.1898 0.8750
Historically black college or university in zip 0.7357 0.2103 0.0000
Constant -1.3642 0.2047 0.0000
Female recruits
83
Table 12. Model results for predicting the number of female recruits by zip code
Variable Coefficient Standard Error p-value
Distance to NRS -0.0234 0.0012 0.0000
Distance to NRS squared 0.0000 0.0000 0.0000
QMA 17-19 I-IIIA black 0.0062 0.0003 0.0000
QMA 17-19 I-IIIA Hispanic 0.0043 0.0003 0.0000
QMA 17-19 I-IIIA white 0.0006 0.0000 0.0000
Navy Awareness Index in 2009 0.3339 0.1113 0.0030
USAF recruiters in 2009 0.1806 0.0390 0.0000
USA recruiters 0.0139 0.0063 0.0270
USMC recruiters 0.1435 0.0199 0.0000
USN recruiters 0.0545 0.0082 0.0000
Violent crime 0.0003 0.0001 0.0340
Property crime 0.0003 0.0000 0.0000
Veteran population 17-44 0.2730 0.0547 0.0000
Veteran population 45-64 0.2248 0.0584 0.0000
Veteran population 65-84 -0.2158 0.0617 0.0000
Veteran population 85 and up 0.5422 0.1876 0.0040
Constant -2.5881 0.1348 0.0000
------------- ----------- ----------- --------
Non-zero model
Distance to closest college / university 0.1672 0.0183 0.0000
Distance squared -0.0023 0.0003 0.0000
College or university in zip -0.0297 0.1694 0.8610
Size of the college / university 0.0338 0.0458 0.4610
Size by distance interaction -0.0053 0.0032 0.0980
Multiple school flag -0.1497 0.2894 0.6050
Historically black college or university in zip -0.0822 0.2257 0.7160
Constant -1.6590 0.2308 0.0000
84
Table 13. Model results for predicting the number of female A-cell recruits by zip code
Variable Coefficient Standard Error p-value
Distance to NRS -0.0262 0.0014 0.0000
Distance to NRS squared 0.0000 0.0000 0.0000
QMA 17-19 I-IIIA black 0.0062 0.0003 0.0000
QMA 17-19 I-IIIA Hispanic 0.0042 0.0003 0.0000
QMA 17-19 I-IIIA white 0.0007 0.0000 0.0000
Navy Awareness Index in 2009 0.5327 0.1245 0.0000
USAF recruiters in 2009 0.1747 0.0441 0.0000
USA recruiters 0.0156 0.0071 0.0290
USMC recruiters 0.1556 0.0222 0.0000
USN recruiters 0.0570 0.0091 0.0000
Violent crime 0.0004 0.0001 0.0060
Property crime 0.0002 0.0000 0.0000
Veteran population 17-44 0.2659 0.0604 0.0000
Veteran population 45-64 0.2334 0.0645 0.0000
Veteran population 65-84 -0.1608 0.0675 0.0170
Veteran population 85 and up 0.5576 0.2024 0.0060
Constant -3.0119 0.1519 0.0000
------------- ----------- ----------- --------
Non-zero model
Distance to closest college / university 0.1819 0.0229 0.0000
Distance squared -0.0025 0.0004 0.0000
College or university in zip 0.0856 0.1901 0.6520
Size of the college / university 0.0656 0.0527 0.2130
Size by distance interaction -0.0079 0.0038 0.0400
Multiple school flag 0.1247 0.2783 0.6540
Historically black college or university in zip 0.0312 0.2533 0.9020
Constant -1.8709 0.2796 0.0000
85
This page intentionally left blank.
86
Appendix C: Examples of data sources for
medical officer goaling
In table 14, we present the total number of graduates by medical
school and gender for 2008 through 2010. Table 15 displays the
number of graduates in osteopathic medicine.
87
Mercer 32 21 53 27 36 63 27 33 60
Morehouse 31 20 51 34 22 56 28 16 44
HI Hawaii-Burns 37 21 58 35 25 60 28 30 58
IA Iowa-Carver 58 78 136 66 78 144 69 62 131
IL Chicago Med-Franklin 85 98 183 89 93 182 78 108 186
Chicago-Pritzker 59 39 98 51 62 113 55 60 115
Illinois 153 155 308 143 157 300 160 170 330
Loyola-Stritch 74 65 139 64 68 132 69 63 132
Northwestern-Feinberg 75 91 166 75 92 167 77 78 155
Rush 77 48 125 60 62 122 73 64 137
Southern Illinois 37 36 73 35 30 65 36 33 69
IN Indiana 118 142 260 120 147 267 116 148 264
KS Kansas 69 94 163 79 76 155 79 89 168
KY Kentucky 36 59 95 46 49 95 41 55 96
Louisville 57 89 146 52 84 136 54 84 138
LA LSU New Orleans 76 79 155 72 98 170 64 101 165
LSU Shreveport 44 53 97 49 61 110 51 61 112
Tulane 65 90 155 60 72 132 66 98 164
MA Boston 81 71 152 90 64 154 86 67 153
Harvard 88 86 174 85 90 175 78 75 153
Massachusetts 53 49 102 50 49 99 61 40 101
Tufts 78 96 174 78 91 169 70 99 169
MD Johns Hopkins 57 44 101 52 73 125 58 60 118
Maryland 85 61 146 76 64 140 98 62 160
Uniformed Services-Hebert 49 108 157 45 113 158 40 123 163
MI Michigan 74 95 169 73 88 161 85 80 165
Michigan State 51 47 98 52 38 90 65 52 117
Wayne State 115 131 246 117 135 252 144 129 273
MN Mayo 17 19 36 20 19 39 22 10 32
Minnesota 110 98 208 94 110 204 103 103 206
MO Missouri Columbia 44 45 89 46 39 85 46 54 100
Missouri Kansas City 58 27 85 51 38 89 50 40 90
St Louis 71 79 150 61 90 151 65 98 163
Washington Univ. St Louis 58 61 119 57 57 114 52 63 115
MS Mississippi 48 50 98 29 65 94 55 60 115
NC Duke 37 53 90 53 54 107 49 56 105
East Carolina-Brody 36 36 72 32 35 67 33 34 67
North Carolina 78 87 165 78 79 157 65 75 140
Wake Forest 49 55 104 49 57 106 50 65 115
ND North Dakota 29 31 60 29 30 59 28 28 56
NE Creighton 55 66 121 62 67 129 59 58 117
Nebraska 52 63 115 50 65 115 51 69 120
88
NH Dartmouth 33 29 62 35 28 63 40 46 86
NJ UMDNJ New Jersey 70 76 146 84 78 162 78 90 168
UMDNJ-RW Johnson 85 69 154 80 70 150 84 67 151
NM New Mexico 43 27 70 34 35 69 42 35 77
NV Nevada 27 22 49 25 26 51 27 28 55
NY Albany 77 56 133 87 51 138 67 60 127
Buffalo 66 66 132 88 61 149 58 71 129
Columbia 59 75 134 75 83 158 77 91 168
Cornell-Weill 45 51 96 52 40 92 45 53 98
Einstein 88 86 174 97 82 179 98 80 178
Mount Sinai 63 55 118 63 58 121 68 48 116
New York Medical 96 96 192 101 82 183 104 88 192
New York University 77 77 154 81 83 164 85 90 175
Rochester 57 34 91 47 42 89 44 50 94
SUNY Downstate 107 94 201 93 106 199 80 93 173
SUNY Upstate 75 68 143 70 84 154 80 70 150
Stony Brook 62 42 104 57 51 108 49 64 113
OH Case Western 65 70 135 75 93 168 71 89 160
Cincinnati 67 89 156 65 78 143 66 96 162
Northeastern Ohio 58 43 101 56 64 120 58 52 110
Ohio State 73 129 202 66 130 196 86 115 201
Toledo 68 72 140 57 78 135 56 85 141
Wright State-Boonshoft 52 37 89 56 42 98 56 34 90
OK Oklahoma 60 79 139 52 96 148 60 89 149
OR Oregon 64 39 103 59 62 121 66 61 127
PA Drexel 127 117 244 126 107 233 131 129 260
Jefferson 111 104 215 128 129 257 129 115 244
Penn State 61 66 127 67 66 133 71 77 148
Pennsylvania 70 71 141 72 79 151 84 72 156
Pittsburgh 67 71 138 73 74 147 56 75 131
Temple 85 96 181 80 83 163 74 87 161
PR Caribe 25 25 50 31 27 58 33 28 61
Ponce 38 35 73 32 34 66 29 29 58
Puerto Rico 53 47 100 42 50 92 47 47 94
San Juan Bautista 41 11 52 47 29 76 33 26 59
RI Brown-Alpert 42 28 70 54 36 90 45 52 97
SC MU South Carolina 64 72 136 60 70 130 66 72 138
South Carolina 38 38 76 35 33 68 30 48 78
SD South Dakota-Sanford 22 31 53 22 25 47 21 25 46
TN East Tennessee-Quillen 26 30 56 35 27 62 28 34 62
Meharry 51 44 95 35 34 69 59 26 85
Tennessee 56 96 152 49 89 138 55 90 145
89
Vanderbilt 36 58 94 46 57 103 51 67 118
TX Baylor 81 98 179 74 83 157 71 81 152
Texas A&M 35 41 76 39 38 77 41 34 75
Texas Tech 57 79 136 61 66 127 64 79 143
UT Galveston 103 94 197 103 95 198 108 113 221
UT HSC San Antonio 128 74 202 108 91 199 122 89 211
UT Houston 86 102 188 84 105 189 94 128 222
UT Southwestern 94 127 221 113 120 233 87 117 204
UT Utah 38 60 98 40 59 99 35 64 99
VA Eastern Virginia 55 54 109 57 49 106 43 56 99
Virginia 57 73 130 64 77 141 69 72 141
Virginia Commonwealth 86 95 181 82 92 174 95 92 187
VT Vermont 57 24 81 57 48 105 61 47 108
WA U Washington 95 74 169 89 91 180 89 80 169
WI MC Wisconsin 87 99 186 89 112 201 99 95 194
Wisconsin 64 71 135 99 75 174 70 71 141
WV Marshall-Edwards 15 27 42 21 31 52 23 38 61
West Virginia 38 50 88 38 60 98 37 66 103
90
Table 15. Graduates in osteopathic medicine
Academic
Year College Total Graduates Male Female % Female
2009-10 ATSU-KCOM 165 101 64 39%
2009-10 AZCOM/MWU 137 99 38 28%
2009-10 CCOM/MWU 171 85 86 50%
2009-10 DMU-COM 207 125 82 40%
2009-10 GA-PCOM 66 29 37 56%
2009-10 KCUMB-COM 239 119 120 50%
2009-10 LECOM 215 120 95 44%
2009-10 LECOM Bradenton 142 69 73 51%
2009-10 MSUCOM 187 89 98 52%
2009-10 NSU-COM 215 107 108 50%
2009-10 NYCOM/NYIT 269 118 151 56%
2009-10 OSU-COM 82 43 39 48%
2009-10 OU-COM 110 55 55 50%
2009-10 PCOM 235 103 132 56%
2009-10 PCSOM 66 33 33 50%
2009-10 TUCOM-CA 129 68 61 47%
2009-10 TUNCOM 120 75 45 38%
2009-10 UMDNJ-SOM 100 46 54 54%
2009-10 UNE-COM 112 56 56 50%
2009-10 UNTHSC/TCOM 151 83 68 45%
2009-10 VCOM 149 63 86 58%
2009-10 Western U/COMP 203 104 99 49%
2009-10 WVSOM 161 72 89 55%
2008-09 ATSU-KCOM 166 104 62 37%
2008-09 AZCOM/MWU 149 83 66 44%
2008-09 CCOM/MWU 163 68 95 58%
2008-09 DMU-COM 197 95 102 52%
2008-09 GA-PCOM 73 35 38 52%
2008-09 KCUMB-COM 234 114 120 51%
2008-09 LECOM 217 119 98 45%
2008-09 LECOM Bradenton 159 82 77 48%
2008-09 MSUCOM 198 87 111 56%
2008-09 NSU-COM 218 105 113 52%
2008-09 NYCOM/NYIT 290 138 152 52%
2008-09 OSU-COM 82 41 41 50%
2008-09 OU-COM 103 44 59 57%
2008-09 PCOM 269 126 143 53%
2008-09 PCSOM 74 42 32 43%
91
2008-09 TUCOM-CA 133 51 82 62%
2008-09 TUNCOM 91 47 44 48%
2008-09 UMDNJ-SOM 92 35 57 62%
2008-09 UNE-COM 116 54 62 53%
2008-09 UNTHSC/TCOM 128 62 66 52%
2008-09 VCOM 139 73 66 47%
2008-09 Western U/COMP 196 98 98 50%
2008-09 WVSOM 101 60 41 41%
92
Appendix D: Review of incentives literature
Our sponsor asked us to briefly review literature on incentives; we
include our review in this appendix. For additional reading, we also
suggest the following references: Cooke [19], Cooke [20],
Samuelson et al. [21], and Jehn and Shughart [22].
Background
After the draft ended in 1973, each of the services created incentive
programs to motivate recruiters. While there are differences among
these programs, there are also many commonalities. For example,
all the services have Recruiter of the Year, Quarter, and Month
awards, as well as awards at different levels of command [23].
Recruiter incentives
In a 2001 study, Emerson analyzed factors that influenced Navy re-
cruiter motivations to meet recruiting objectives. Data for the study
were collected through an online survey given to the enlisted re-
93
cruiting force. The survey was meant to identify recruiters’ attitudes
about certain incentives. Sabbaticals and financial awards ranked
highest on the list of extrinsic tangible incentives, but data analysis
indicated that recruiters rank intangible incentives even higher.
The top two intangible incentives were: (1) wanting to avoid letting
down their station and (2) feeling good for meeting the mission
[23].
Loving also discovered that a good command climate has the largest
impact on recruiter motivation [24]. Two other studies reveal that
how leaders choose to motivate, set goals for, and discipline recruit-
ers can affect recruiter performance (see [23 and 25]). Incentive
preferences differed by paygrade, volunteer status, and membership
in the Career Recruiting Force.
Geographic location
Incentive program
The study found that recruiters are more productive when recruit-
ing people who are like themselves. For example, women were bet-
ter at recruiting women [26]. This finding has potential policy
implications, particularly when services must target populations with
specific attributes. It may be beneficial to increase the number of
94
recruiters with a certain profile (e.g., male Hispanics) if, and when,
a service aims to recruit more male Hispanics. Similarly, one could
infer that some involvement of a surgeon might be helpful in re-
cruiting another surgeon. Clearly, this is difficult, in practice, when
it comes to recruiting for critical jobs in the medical field. One solu-
tion might be to have nonmedical recruiters put forth most of the
effort needed to recruit for critical medical jobs, but have some rep-
resentatives of those fields come in and give seminars on their jobs
with some regularity.
95
program incentivizes them to sign up high school graduates who
can attend recruit training within 30 to 60 days. This kind of cam-
paign is usually highly publicized, with recruiter performance dis-
seminated daily, to spur competition among Marines. One year
there was a baseball theme to the program, and awards included
baseballs and bats. More substantial rewards, such as meritorious
promotions, went to the most productive recruiters. This is further
evidence that spurring competition and using awards could make a
difference in medical recruiting as well.
96
References
[1] Nonresident Training Course. Navy Counselor 1 & C (Recruiter).
NAVEDTRA 14172. Jan. 1995.
[7] Don Bohn and Edward Schmitz. A Zip Code Based Production
Function. United States Navy Recruiting Command. Dec. 1992.
97
10] Jennifer Gibson et al. Zip Code Valuation Study Technical
Report: Predicting Navy Accessions. Defense Human Re-
source Activity, JAMRS Report No. 2009-16. Dec. 2009.
[15] Jennie Wenger and Laura Kelley. Marine Corps Officer Re-
cruiting: Which Schools Did Officers Attend? CNA Research
Memorandum D0014599.A2/Final. Sep. 2006.
98
[19] Timothy W. Cooke. Individual Incentives in Navy Recruiting.
CNA Research Memorandum 86-289. Dec. 1986.
99
[29] Samuelson et al, Productivity Effects of Changes in the Size
of the Enlisted Recruiter Force. CNA Research Memoran-
dum D0013975.A2/Final, May 2006
100
Bibliography
Asch, Beth. Navy Recruiter Productivity and the Freeman Plan. RAND
Corporation Report R-3713-FMP. 1990.
James E. Parco and David A. Levy, eds. Attitudes Aren’t Free: Thinking
Deeply about Diversity in the US Armed Forces. Maxwell Air Force
Base, AL: Air University Press, 2010.
101
This page intentionally left blank.w
102
List of tables
Table 1. Results from the Any Contracts Model: Coefficients for
modeling productive zip codes ............................................. 24
Table 7. Model results for predicting A-cell recruits by zip code ....... 77
Table 13. Model results for predicting the number of female A-cell
recruits by zip code ................................................................ 85
Table 14. Total graduates by U.S. medical school and gender ............ 87
103
This page intentionally ;left blank.
104
CRM D0026005.A2/Final