← Back to Projects

Discretionary Trees: Understanding Street-Level Bureaucracy via Machine Learning

Authors: Gaurab Pokharel¹, Patrick J. Fowler², and Sanmay Das¹

Affiliations: ¹Virginia Tech, ²Washington University in St. Louis

Venue: Proceedings of the AAAI Conference on Artificial Intelligence (AAAI 2024)


Overview

Street-level bureaucrats (caseworkers, triage nurses, etc.) implement policy in real time while exercising judgment in individual cases. That discretion can be valuable—allowing exceptions when context matters—but it can also introduce inequities if applied inconsistently.

This project uses machine learning to answer three practical questions:

  1. Simple rules: How much of real decision-making can be summarized as short, human-usable heuristics?
  2. Consistency: When decisions go beyond simple rules, are they still procedurally consistent?
  3. Discretion effects: When discretion happens, who gets it—and does it appear outcome-improving?

Data and Setting

We use administrative data from the St. Louis Homeless Management Information System (HMIS), covering household characteristics and assigned interventions during 2007–2014, when assignment was not fully formulaic.

We focus primarily on the tradeoff between:

  • Emergency Shelter (ES) (less intensive)
  • Transitional Housing (TH) (more intensive)

Method

1) “Default heuristics” via short decision trees

We learn short, constrained decision trees that approximate the kind of quick heuristics a caseworker could plausibly apply in “easy” cases.

SERDT Trees
Figure 1. The decision trees generated by SER-DT for the assignment of Emergency Shelter (left) and Transitional Housing (right). The left branches of each node indicate that the parent condition was satisfied, whereas the right branch indicates the condition was false. The generated trees are small but can encompass a significant portion of the correctly made decisions.

2) Consistency via a high-capacity model

We train a stronger predictive model (e.g., gradient-boosted trees) to test whether caseworker decisions are consistent even when not captured by the short heuristics.

XGBoost Performance
Figure 2. ROC curves for each tree used to learn intervention assignments using the XGBoost algorithm. Predictive discrimination, measured using the area under the ROC curve is very high, implying significant consistency in caseworker decision-making.

3) Operationalizing discretion

We treat discretionary decisions as those where the actual assignment differs from what the short-tree heuristic would have predicted—capturing “beyond-heuristic” judgment.


Key Findings

A) Many decisions are explainable by simple rules

Short trees capture a meaningful share of assignment decisions, consistent with the idea that caseworkers rely on heuristics for many cases.

B) Beyond heuristics, decisions are still highly consistent

A more flexible model predicts assignments extremely well, suggesting caseworkers are procedurally consistent rather than arbitrary.

C) Discretion is targeted and not random

When discretion occurs, it tends to be applied to lower-vulnerability households rather than the most vulnerable.

Nonparametric test on vulnerability
Figure 3. Non-parametric test comparing vulnerability scores between the discretionary subgroups and the population predicted to receive the intervention.

D) Discretionary upgrades align with higher marginal benefit

Discretionary moves from predicted ES → actual TH are associated with higher expected marginal benefit than random reassignment, while discretionary TH → ES does not show a comparable loss.

Nonparametric test on marginal utility
Figure 4. Non-parametric test comparing the marginal gain in utility resulting from the switch in interventions between the discretionary subgroups and the population predicted to receive the intervention.

E) A compact way to visualize “discretion” in the dataset

A nice at-a-glance graphic is the predicted-vs-actual distribution (the “where mismatches happen” plot).

Nonparametric test on marginal utility
Figure 5. Distribution of the number of records for predicted versus actual assignment of intervention. We focus our analysis of discretion quantification in the subgroups marked within the red box.

Why this matters for decision support

This is evidence for a “human-in-the-loop” design story: simple rules explain a large portion of decisions, and discretion appears to be applied in structured ways that (at least on average) align with improved outcomes for certain cases. That motivates decision support tools that make heuristics explicit and support/monitor discretion, rather than replacing frontline judgment.


Citation

@article{Pokharel_Das_Fowler_2024, 
    title={Discretionary Trees: Understanding Street-Level Bureaucracy via Machine Learning}, 
    volume={38},
    ISSN={2374-3468, 2159-5399}, 
    url={https://ojs.aaai.org/index.php/AAAI/article/view/30236}, 
    DOI={10.1609/aaai.v38i20.30236},  
    number={20}, 
    journal={Proceedings of the AAAI Conference on Artificial Intelligence}, 
    author={Pokharel, Gaurab and Das, Sanmay and Fowler, Patrick}, 
    year={2024}, 
    month=mar, 
    pages={22303–22312} 
  }

This work was partially supported by the NSF (IIS-1939677, IIS-2127752) and by Amazon through an NSF FAI award.


Paper PDF