2018
DOI: 10.1613/jair.1.11233
|View full text |Cite
|
Sign up to set email alerts
|

Human-Machine Collaborative Optimization via Apprenticeship Scheduling

Abstract: Coordinating agents to complete a set of tasks with intercoupled temporal and resource constraints is computationally challenging, yet human domain experts can solve these difficult scheduling problems using paradigms learned through years of apprenticeship. A process for manually codifying this domain knowledge within a computational framework is necessary to scale beyond the "single-expert, single-trainee" apprenticeship model. However, human domain experts often have difficulty describing their decision-mak… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
15
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 23 publications
(20 citation statements)
references
References 82 publications
0
15
0
Order By: Relevance
“…In robotics, preference learning typically involves learning a human's preference and adapting a robot's behavior to better collaborate with human-beings [16,28]. There are several methods that could infer preference information without asking the human to explicitly provide preference information [14,15]. For example, Schafer et al [34] utilizes collaborative filtering for a recommender system and Xu et al [42] formulates a meta-IRL problem that could learn a prior over preferences.…”
Section: Related Workmentioning
confidence: 99%
“…In robotics, preference learning typically involves learning a human's preference and adapting a robot's behavior to better collaborate with human-beings [16,28]. There are several methods that could infer preference information without asking the human to explicitly provide preference information [14,15]. For example, Schafer et al [34] utilizes collaborative filtering for a recommender system and Xu et al [42] formulates a meta-IRL problem that could learn a prior over preferences.…”
Section: Related Workmentioning
confidence: 99%
“…The framework we consider is common to military weapon target engagements (Aviv & Kress, 1997;Gombolay et al, 2018;Kisi, 1976;Mastran & Thomas, 1973), shooting problems (Glazebrook & Washburn, 2004;Sato, 1997a), and more broadly to stochastic sequential resource allocation (Sato, 1996(Sato, , 1997b. Monotonicity properties for the related bomber problem, where a bomber has to survive sequential engagements with enemy aircraft by firing a volley (one or more) of weapons are discussed in (Weber, 2013).…”
Section: Prior Workmentioning
confidence: 99%
“…For example, when a malfunction occurs or when operating conditions require a new schedule, replanning needs to be executed promptly as machine idle time can be extremely costly. (e.g., on the order of $10,000 per minute for some applications [12]). To address this issue, system operators typically seek approximate solutions to the original scheduling problems.…”
Section: Introductionmentioning
confidence: 99%