3 * \brief Process management service.
7 * Copyright (c) 2017, ETH Zurich.
10 * This file is distributed under the terms in the attached LICENSE file.
11 * If you do not find this file, copies can be found by writing to:
12 * ETH Zurich D-INFK, Haldeneggsteig 4, CH-8092 Zurich. Attn: Systems Group.
15 #include <barrelfish/barrelfish.h>
16 #include <barrelfish/nameservice_client.h>
17 #include <barrelfish/spawn_client.h>
18 #include <if/monitor_defs.h>
19 #include <if/proc_mgmt_defs.h>
20 #include <if/spawn_defs.h>
24 #include "pending_clients.h"
25 #include "spawnd_state.h"
28 * \brief Handler for message add_spawnd, for the local monitor binding.
30 static void add_spawnd_handler(struct proc_mgmt_binding *b, coreid_t core_id,
33 if (spawnd_state_exists(core_id)) {
34 DEBUG_ERR(PROC_MGMT_ERR_SPAWND_EXISTS, "spawnd_state_exists");
38 // Bind with the spawnd.
39 struct spawn_binding *spawnb;
40 errval_t err = spawn_bind_iref(iref, &spawnb);
41 if (err_is_fail(err)) {
42 DEBUG_ERR(err, "spawn_bind_iref");
46 err = spawnd_state_alloc(core_id, spawnb);
47 if (err_is_fail(err)) {
48 DEBUG_ERR(err, "spawnd_state_alloc");
51 debug_printf("Process manager bound with spawnd.%u on iref %u\n", core_id,
56 * \brief Handler for message add_spawnd, for non-monitor bindings.
58 static void add_spawnd_handler_non_monitor(struct proc_mgmt_binding *b,
59 coreid_t core_id, iref_t iref)
61 // debug_printf("Ignoring add_spawnd call: %s\n",
62 // err_getstring(PROC_MGMT_ERR_NOT_MONITOR));
65 static bool cleanup_request_sender(struct msg_queue_elem *m);
68 * General-purpose handler for replies from spawnd.
70 static void spawn_reply_handler(struct spawn_binding *b, errval_t spawn_err)
72 struct pending_client *cl =
73 (struct pending_client*) spawnd_state_dequeue_recv(b->st);
75 struct pending_spawn *spawn = NULL;
76 struct pending_span *span = NULL;
77 struct pending_kill_cleanup *kc = NULL;
79 struct domain_entry *entry;
81 errval_t err, resp_err;
84 case ClientType_Spawn:
85 case ClientType_SpawnWithCaps:
86 spawn = (struct pending_spawn*) cl->st;
88 if (err_is_ok(spawn_err)) {
89 err = domain_spawn(spawn->cap_node, spawn->core_id);
90 if (cl->type == ClientType_Spawn) {
91 resp_err = cl->b->tx_vtbl.spawn_response(cl->b, NOP_CONT,
92 err, spawn->cap_node->domain_cap);
94 resp_err = cl->b->tx_vtbl.spawn_with_caps_response(cl->b,
95 NOP_CONT, err, spawn->cap_node->domain_cap);
102 case ClientType_Span:
103 span = (struct pending_span*) cl->st;
105 if (entry->status == DOMAIN_STATUS_RUNNING) {
106 resp_err = cl->b->tx_vtbl.span_response(cl->b, NOP_CONT,
113 case ClientType_Cleanup:
114 kc = (struct pending_kill_cleanup*) cl->st;
117 assert(entry->num_spawnds_resources > 0);
118 assert(entry->status != DOMAIN_STATUS_CLEANED);
120 --entry->num_spawnds_resources;
121 if (entry->num_spawnds_resources == 0) {
122 entry->status = DOMAIN_STATUS_CLEANED;
124 // At this point, the domain exists in state CLEANED for history
125 // reasons. For instance, if some other domain issues a wait
126 // call for this one, the process manager can return the exit
127 // status directly. At some point, however, we might want to
128 // just clean up the domain entry and recycle the domain cap.
134 case ClientType_Kill:
135 case ClientType_Exit:
136 kc = (struct pending_kill_cleanup*) cl->st;
139 assert(entry->num_spawnds_running > 0);
140 assert(entry->status != DOMAIN_STATUS_STOPPED);
142 --entry->num_spawnds_running;
144 if (entry->num_spawnds_running == 0) {
145 entry->status = DOMAIN_STATUS_STOPPED;
147 if (cl->type == ClientType_Kill) {
148 entry->exit_status = EXIT_STATUS_KILLED;
149 resp_err = cl->b->tx_vtbl.kill_response(cl->b, NOP_CONT,
153 struct domain_waiter *waiter = entry->waiters;
154 while (waiter != NULL) {
155 waiter->b->tx_vtbl.wait_response(waiter->b, NOP_CONT,
158 struct domain_waiter *tmp = waiter;
159 waiter = waiter->next;
163 for (coreid_t i = 0; i < MAX_COREID; ++i) {
164 if (entry->spawnds[i] == NULL) {
168 struct spawn_binding *spb = entry->spawnds[i]->b;
170 struct pending_kill_cleanup *cleanup =
171 (struct pending_kill_cleanup*) malloc(
172 sizeof(struct pending_kill_cleanup));
174 cleanup->domain_cap = kc->domain_cap;
175 cleanup->entry = entry;
177 struct pending_client *cleanup_cl =
178 (struct pending_client*) malloc(
179 sizeof(struct pending_client));
180 cleanup_cl->b = cl->b;
181 cleanup_cl->type = ClientType_Cleanup;
182 cleanup_cl->st = cleanup;
184 struct msg_queue_elem *msg = (struct msg_queue_elem*) malloc(
185 sizeof(struct msg_queue_elem));
186 msg->st = cleanup_cl;
187 msg->cont = cleanup_request_sender;
189 err = spawnd_state_enqueue_send(entry->spawnds[i], msg);
191 if (err_is_fail(err)) {
192 DEBUG_ERR(err, "enqueuing cleanup request");
204 USER_PANIC("Unknown client type in spawn_reply_handler: %u\n",
212 * \brief Handler for sending spawn requests.
214 static bool spawn_request_sender(struct msg_queue_elem *m)
216 struct pending_client *cl = (struct pending_client*) m->st;
217 struct pending_spawn *spawn = (struct pending_spawn*) cl->st;
218 spawn->b->rx_vtbl.spawn_reply = spawn_reply_handler;
221 bool with_caps = !(capref_is_null(spawn->inheritcn_cap) &&
222 capref_is_null(spawn->argcn_cap));
224 err = spawn->b->tx_vtbl.spawn_with_caps_request(spawn->b, NOP_CONT,
226 spawn->cap_node->domain_cap,
232 spawn->inheritcn_cap,
236 err = spawn->b->tx_vtbl.spawn_request(spawn->b, NOP_CONT, cap_procmng,
237 spawn->cap_node->domain_cap,
238 spawn->path, spawn->argvbuf,
239 spawn->argvbytes, spawn->envbuf,
240 spawn->envbytes, spawn->flags);
243 if (err_is_fail(err)) {
244 if (err_no(err) == FLOUNDER_ERR_TX_BUSY) {
247 USER_PANIC_ERR(err, "sending spawn request");
257 * \brief Handler for sending span requests.
259 static bool span_request_sender(struct msg_queue_elem *m)
261 struct pending_client *cl = (struct pending_client*) m->st;
262 struct pending_span *span = (struct pending_span*) cl->st;
265 span->b->rx_vtbl.spawn_reply = spawn_reply_handler;
266 err = span->b->tx_vtbl.span_request(span->b, NOP_CONT, cap_procmng,
267 span->domain_cap, span->vroot,
270 if (err_is_fail(err)) {
271 if (err_no(err) == FLOUNDER_ERR_TX_BUSY) {
274 USER_PANIC_ERR(err, "sending span request");
284 * \brief Handler for sending kill requests.
286 static bool kill_request_sender(struct msg_queue_elem *m)
288 struct pending_client *cl = (struct pending_client*) m->st;
289 struct pending_kill_cleanup *kill = (struct pending_kill_cleanup*) cl->st;
292 kill->b->rx_vtbl.spawn_reply = spawn_reply_handler;
293 err = kill->b->tx_vtbl.kill_request(kill->b, NOP_CONT, cap_procmng,
296 if (err_is_fail(err)) {
297 if (err_no(err) == FLOUNDER_ERR_TX_BUSY) {
300 USER_PANIC_ERR(err, "sending kill request");
310 * \brief Handler for sending cleanup requests.
312 static bool cleanup_request_sender(struct msg_queue_elem *m)
314 struct pending_client *cl = (struct pending_client*) m->st;
315 struct pending_kill_cleanup *cleanup = (struct pending_kill_cleanup*) cl->st;
318 cleanup->b->rx_vtbl.spawn_reply = spawn_reply_handler;
319 err = cleanup->b->tx_vtbl.cleanup_request(cleanup->b, NOP_CONT,
321 cleanup->domain_cap);
323 if (err_is_fail(err)) {
324 if (err_no(err) == FLOUNDER_ERR_TX_BUSY) {
327 USER_PANIC_ERR(err, "sending cleanup request");
337 * \brief Common bits of the spawn and spawn_with_caps handlers.
339 static errval_t spawn_handler_common(struct proc_mgmt_binding *b,
340 enum ClientType type,
341 coreid_t core_id, const char *path,
342 const char *argvbuf, size_t argvbytes,
343 const char *envbuf, size_t envbytes,
344 struct capref inheritcn_cap,
345 struct capref argcn_cap, uint8_t flags)
347 if (!spawnd_state_exists(core_id)) {
348 return PROC_MGMT_ERR_INVALID_SPAWND;
351 struct spawnd_state *spawnd = spawnd_state_get(core_id);
352 assert(spawnd != NULL);
353 struct spawn_binding *cl = spawnd->b;
357 if (domain_should_refill_caps()) {
358 err = domain_prealloc_caps();
359 if (err_is_fail(err)) {
360 return err_push(err, PROC_MGMT_ERR_CREATE_DOMAIN_CAP);
364 struct domain_cap_node *cap_node = next_cap_node();
366 struct pending_spawn *spawn = (struct pending_spawn*) malloc(
367 sizeof(struct pending_spawn));
368 spawn->cap_node = cap_node;
369 // spawn->domain_cap = domain_cap;
371 spawn->core_id = core_id;
373 spawn->argvbuf = argvbuf;
374 spawn->argvbytes = argvbytes;
375 spawn->envbuf = envbuf;
376 spawn->envbytes = envbytes;
377 spawn->inheritcn_cap = inheritcn_cap;
378 spawn->argcn_cap = argcn_cap;
379 spawn->flags = flags;
381 struct pending_client *spawn_cl = (struct pending_client*) malloc(
382 sizeof(struct pending_client));
384 spawn_cl->type = type;
385 spawn_cl->st = spawn;
387 struct msg_queue_elem *msg = (struct msg_queue_elem*) malloc(
388 sizeof(struct msg_queue_elem));
390 msg->cont = spawn_request_sender;
392 err = spawnd_state_enqueue_send(spawnd, msg);
393 if (err_is_fail(err)) {
394 DEBUG_ERR(err, "enqueuing spawn request");
404 * \brief Handler for rpc spawn.
406 static void spawn_handler(struct proc_mgmt_binding *b, coreid_t core_id,
407 const char *path, const char *argvbuf,
408 size_t argvbytes, const char *envbuf, size_t envbytes,
411 errval_t err, resp_err;
412 err = spawn_handler_common(b, ClientType_Spawn, core_id, path, argvbuf,
413 argvbytes, envbuf, envbytes, NULL_CAP, NULL_CAP,
416 if (err_is_fail(err)) {
417 resp_err = b->tx_vtbl.spawn_response(b, NOP_CONT, err, NULL_CAP);
418 if (err_is_fail(resp_err)) {
419 DEBUG_ERR(resp_err, "failed to send spawn_response");
425 * \brief Handler for rpc spawn_with_caps.
427 static void spawn_with_caps_handler(struct proc_mgmt_binding *b,
428 coreid_t core_id, const char *path,
429 const char *argvbuf, size_t argvbytes,
430 const char *envbuf, size_t envbytes,
431 struct capref inheritcn_cap,
432 struct capref argcn_cap, uint8_t flags)
434 errval_t err, resp_err;
435 err = spawn_handler_common(b, ClientType_SpawnWithCaps, core_id, path,
436 argvbuf, argvbytes, envbuf, envbytes,
437 inheritcn_cap, argcn_cap, flags);
438 if (err_is_ok(err)) {
439 // Will respond to client when we get the reply from spawnd.
443 resp_err = b->tx_vtbl.spawn_with_caps_response(b, NOP_CONT, err,
445 if (err_is_fail(resp_err)) {
446 DEBUG_ERR(resp_err, "failed to send spawn_with_caps_response");
451 * \brief Handler for rpc span.
453 static void span_handler(struct proc_mgmt_binding *b, struct capref domain_cap,
454 coreid_t core_id, struct capref vroot,
455 struct capref dispframe)
457 errval_t err, resp_err;
458 struct domain_entry *entry = NULL;
459 err = domain_get_by_cap(domain_cap, &entry);
460 if (err_is_fail(err)) {
461 goto respond_with_err;
464 assert(entry != NULL);
465 if (entry->status != DOMAIN_STATUS_RUNNING) {
466 err = PROC_MGMT_ERR_DOMAIN_NOT_RUNNING;
467 goto respond_with_err;
470 if (entry->spawnds[core_id] != NULL) {
471 // TODO(razvan): Maybe we want to allow the same domain to span multiple
472 // dispatchers onto the same core?
473 err = PROC_MGMT_ERR_ALREADY_SPANNED;
474 goto respond_with_err;
477 if (!spawnd_state_exists(core_id)) {
478 err = PROC_MGMT_ERR_INVALID_SPAWND;
479 goto respond_with_err;
482 struct spawnd_state *spawnd = spawnd_state_get(core_id);
483 assert(spawnd != NULL);
484 struct spawn_binding *cl = spawnd->b;
487 struct pending_span *span = (struct pending_span*) malloc(
488 sizeof(struct pending_span));
489 span->domain_cap = domain_cap;
492 span->core_id = core_id;
494 span->dispframe = dispframe;
496 struct pending_client *span_cl = (struct pending_client*) malloc(
497 sizeof(struct pending_client));
499 span_cl->type = ClientType_Span;
502 struct msg_queue_elem *msg = (struct msg_queue_elem*) malloc(
503 sizeof(struct msg_queue_elem));
505 msg->cont = span_request_sender;
507 err = spawnd_state_enqueue_send(spawnd, msg);
509 if (err_is_fail(err)) {
510 DEBUG_ERR(err, "enqueuing span request");
517 resp_err = b->tx_vtbl.span_response(b, NOP_CONT, err);
518 if (err_is_fail(resp_err)) {
519 DEBUG_ERR(resp_err, "failed to send span_response");
524 * \brief Common bits of the kill and exit handlers.
526 static errval_t kill_handler_common(struct proc_mgmt_binding *b,
527 struct capref domain_cap,
528 enum ClientType type,
531 struct domain_entry *entry;
532 errval_t err = domain_get_by_cap(domain_cap, &entry);
533 if (err_is_fail(err)) {
537 entry->exit_status = exit_status;
538 domain_stop_pending(entry);
540 for (coreid_t i = 0; i < MAX_COREID; ++i) {
541 if (entry->spawnds[i] == NULL) {
545 struct spawn_binding *spb = entry->spawnds[i]->b;
547 struct pending_kill_cleanup *cmd = (struct pending_kill_cleanup*) malloc(
548 sizeof(struct pending_kill_cleanup));
549 cmd->domain_cap = domain_cap;
553 struct pending_client *cl = (struct pending_client*) malloc(
554 sizeof(struct pending_client));
559 struct msg_queue_elem *msg = (struct msg_queue_elem*) malloc(
560 sizeof(struct msg_queue_elem));
562 msg->cont = kill_request_sender;
564 err = spawnd_state_enqueue_send(entry->spawnds[i], msg);
565 if (err_is_fail(err)) {
566 DEBUG_ERR(err, "enqueuing kill request");
577 * \brief Handler for rpc kill.
579 static void kill_handler(struct proc_mgmt_binding *b,
580 struct capref victim_domain_cap)
582 errval_t err = kill_handler_common(b, victim_domain_cap, ClientType_Kill,
584 if (err_is_fail(err)) {
585 errval_t resp_err = b->tx_vtbl.kill_response(b, NOP_CONT, err);
586 if (err_is_fail(resp_err)) {
587 DEBUG_ERR(resp_err, "failed to send kill_response");
593 * \brief Handler for message exit.
595 static void exit_handler(struct proc_mgmt_binding *b, struct capref domain_cap,
598 errval_t err = kill_handler_common(b, domain_cap, ClientType_Exit,
600 if (err_is_fail(err)) {
601 DEBUG_ERR(err, "processing exit_handler for requesting domain, exit "
602 "code %u", exit_status);
604 // Error or not, there's no client to respond to anymore.
608 * \brief Handler for rpc wait.
610 static void wait_handler(struct proc_mgmt_binding *b, struct capref domain_cap)
612 errval_t err, resp_err;
613 struct domain_entry *entry;
614 err = domain_get_by_cap(domain_cap, &entry);
615 if (err_is_fail(err)) {
619 if (entry->status == DOMAIN_STATUS_STOPPED) {
620 // Domain has already been stopped, so just reply with exit status.
624 struct domain_waiter *waiter = (struct domain_waiter*) malloc(
625 sizeof(struct domain_waiter));
627 waiter->next = entry->waiters;
628 entry->waiters = waiter;
629 // Will respond when domain is stopped.
633 resp_err = b->tx_vtbl.wait_response(b, NOP_CONT, err, entry->exit_status);
634 if (err_is_fail(resp_err)) {
635 DEBUG_ERR(resp_err, "failed to send wait_response");
639 static struct proc_mgmt_rx_vtbl monitor_vtbl = {
640 .add_spawnd = add_spawnd_handler,
641 .spawn_call = spawn_handler,
642 .spawn_with_caps_call = spawn_with_caps_handler,
643 .span_call = span_handler,
644 .kill_call = kill_handler,
645 .exit_call = exit_handler,
646 .wait_call = wait_handler
649 static struct proc_mgmt_rx_vtbl non_monitor_vtbl = {
650 .add_spawnd = add_spawnd_handler_non_monitor,
651 .spawn_call = spawn_handler,
652 .spawn_with_caps_call = spawn_with_caps_handler,
653 .span_call = span_handler,
654 .kill_call = kill_handler,
655 .exit_call = exit_handler,
656 .wait_call = wait_handler
660 * \brief Allocates a special LMP endpoint for authenticating with the monitor.
662 static errval_t alloc_ep_for_monitor(struct capref *ep)
664 struct proc_mgmt_lmp_binding *lmpb =
665 malloc(sizeof(struct proc_mgmt_lmp_binding));
666 assert(lmpb != NULL);
668 // setup our end of the binding
669 errval_t err = proc_mgmt_client_lmp_accept(lmpb, get_default_waitset(),
670 DEFAULT_LMP_BUF_WORDS);
671 if (err_is_fail(err)) {
673 return err_push(err, LIB_ERR_PROC_MGMT_CLIENT_ACCEPT);
676 *ep = lmpb->chan.local_cap;
677 lmpb->b.rx_vtbl = monitor_vtbl;
682 static void export_cb(void *st, errval_t err, iref_t iref)
684 if (err_is_fail(err)) {
685 USER_PANIC_ERR(err, "export failed");
688 // Allocate an endpoint for the local monitor, who will use it to inform
689 // us about new spawnd irefs on behalf of other monitors.
691 err = alloc_ep_for_monitor(&ep);
692 if (err_is_fail(err)) {
693 USER_PANIC_ERR(err, "failed to allocate LMP EP for local monitor");
696 // Send the endpoint to the monitor, so it can finish the handshake.
697 struct monitor_binding *mb = get_monitor_binding();
698 err = mb->tx_vtbl.set_proc_mgmt_ep_request(mb, NOP_CONT, ep);
699 if (err_is_fail(err)) {
700 USER_PANIC_ERR(err, "failed to send set_proc_mgmt_ep_request to "
704 // Also register this iref with the name service, for arbitrary client
705 // domains to use for spawn-related ops.
706 err = nameservice_register(SERVICE_BASENAME, iref);
707 if (err_is_fail(err)) {
708 USER_PANIC_ERR(err, "nameservice_register failed");
712 static errval_t connect_cb(void *st, struct proc_mgmt_binding *b)
714 b->rx_vtbl = non_monitor_vtbl;
718 errval_t start_service(void)
720 errval_t err = domain_prealloc_caps();
721 if (err_is_fail(err)) {
722 USER_PANIC_ERR(err_push(err, PROC_MGMT_ERR_CREATE_DOMAIN_CAP),
723 "domain_prealloc_caps in start_service");
726 return proc_mgmt_export(NULL, export_cb, connect_cb, get_default_waitset(),
727 IDC_EXPORT_FLAGS_DEFAULT);