Organize resources into process groups

This commit is contained in:
2026-01-27 14:18:05 +01:00
parent 67b66f2b39
commit 600886a7ee
20 changed files with 387 additions and 457 deletions

View File

@@ -2,5 +2,9 @@ Lock hierarchy for process scheduling:
1. proc_tree_lock
2. cpu->lock
3. proc->lock
4. sq->lock
3. procgroup->lock
4. proc->lock
5. sq->lock
1. procgroup_tree_lock
2. procgroup->lock

View File

@@ -5,29 +5,7 @@
#include <proc/resource.h>
#include <sync/spin_lock.h>
bool proc_create_resource_mem (struct proc_resource_mem* mem, struct proc_resource_mem_init* init) {
if (init->pages == 0)
return false;
if (init->managed) {
mem->paddr = init->paddr;
mem->managed = true;
} else {
uintptr_t paddr = pmm_alloc (init->pages);
if (paddr == PMM_ALLOC_ERR)
return false;
mem->paddr = paddr;
mem->managed = false;
}
mem->pages = mem->alive_pages = init->pages;
return true;
}
void proc_cleanup_resource_mem (struct proc* proc, struct proc_resource* resource) {
(void)proc;
void proc_cleanup_resource_mem (struct proc_resource* resource) {
if (!resource->u.mem.managed)
pmm_free (resource->u.mem.paddr, resource->u.mem.pages);
}

View File

@@ -15,13 +15,6 @@ struct proc_resource_mem {
bool managed;
};
struct proc_resource_mem_init {
uintptr_t paddr;
size_t pages;
bool managed;
};
bool proc_create_resource_mem (struct proc_resource_mem* mem, struct proc_resource_mem_init* init);
void proc_cleanup_resource_mem (struct proc* proc, struct proc_resource* resource);
void proc_cleanup_resource_mem (struct proc_resource* resource);
#endif // _KERNEL_PROC_MEM_H

View File

@@ -85,13 +85,7 @@ static void proc_mutex_resume (struct proc* proc, struct proc_sq_entry* sq_entry
cpu_request_sched (cpu);
}
bool proc_create_resource_mutex (struct proc_mutex* mutex) {
memset (mutex, 0, sizeof (*mutex));
return true;
}
void proc_cleanup_resource_mutex (struct proc* proc, struct proc_resource* resource) {
void proc_cleanup_resource_mutex (struct proc_resource* resource) {
struct proc_mutex* mutex = &resource->u.mutex;
spin_lock_ctx_t ctxmt, ctxsq;

View File

@@ -15,8 +15,7 @@ struct proc_mutex {
struct proc* owner;
};
bool proc_create_resource_mutex (struct proc_mutex* mutex);
void proc_cleanup_resource_mutex (struct proc* proc, struct proc_resource* resource);
void proc_cleanup_resource_mutex (struct proc_resource* resource);
void proc_mutex_lock (struct proc* proc, struct proc_mutex* mutex);
bool proc_mutex_unlock (struct proc* proc, struct proc_mutex* mutex);

View File

@@ -11,6 +11,7 @@
#include <mm/pmm.h>
#include <proc/kpproc_fb.h>
#include <proc/proc.h>
#include <proc/procgroup.h>
#include <proc/resource.h>
#include <rd/rd.h>
#include <sync/rw_spin_lock.h>
@@ -56,16 +57,16 @@ bool proc_map (struct proc* proc, uintptr_t start_paddr, uintptr_t start_vaddr,
flags &= ~(MM_PD_LOCK | MM_PD_RELOAD); /* clear LOCK flag if present, because we lock manualy */
spin_lock (&proc->pd->lock, &ctxprpd);
spin_lock (&proc->procgroup->pd.lock, &ctxprpd);
list_append (proc->mappings, &mapping->proc_mappings_link);
list_append (proc->procgroup->mappings, &mapping->proc_mappings_link);
for (uintptr_t vpage = start_vaddr, ppage = start_paddr; vpage < start_vaddr + pages * PAGE_SIZE;
vpage += PAGE_SIZE, ppage += PAGE_SIZE) {
mm_map_page (proc->pd, ppage, vpage, flags);
mm_map_page (&proc->procgroup->pd, ppage, vpage, flags);
}
spin_unlock (&proc->pd->lock, &ctxprpd);
spin_unlock (&proc->procgroup->pd.lock, &ctxprpd);
return true;
}
@@ -81,9 +82,9 @@ bool proc_unmap (struct proc* proc, uintptr_t start_vaddr, size_t pages) {
if (tail_mapping == NULL)
return false;
spin_lock (&proc->pd->lock, &ctxprpd);
spin_lock (&proc->procgroup->pd.lock, &ctxprpd);
list_foreach (proc->mappings, mapping_link, mapping_link_tmp) {
list_foreach (proc->procgroup->mappings, mapping_link, mapping_link_tmp) {
struct proc_mapping* mapping =
list_entry (mapping_link, struct proc_mapping, proc_mappings_link);
@@ -99,7 +100,7 @@ bool proc_unmap (struct proc* proc, uintptr_t start_vaddr, size_t pages) {
mapping->size = start_vaddr - mapping->vaddr;
list_insert_after (proc->mappings, &mapping->proc_mappings_link,
list_insert_after (proc->procgroup->mappings, &mapping->proc_mappings_link,
&tail_mapping->proc_mappings_link);
used_tail_mapping = true;
@@ -113,7 +114,7 @@ bool proc_unmap (struct proc* proc, uintptr_t start_vaddr, size_t pages) {
} else if ((start_vaddr > mapping->vaddr) && (end_vaddr >= m_end)) { /* shrink right */
mapping->size = start_vaddr - mapping->vaddr;
} else { /* full overlap */
list_remove (proc->mappings, &mapping->proc_mappings_link);
list_remove (proc->procgroup->mappings, &mapping->proc_mappings_link);
free (mapping);
}
}
@@ -123,10 +124,10 @@ bool proc_unmap (struct proc* proc, uintptr_t start_vaddr, size_t pages) {
free (tail_mapping);
for (uintptr_t vpage = start_vaddr; vpage < end_vaddr; vpage += PAGE_SIZE) {
mm_unmap_page (proc->pd, vpage, 0);
mm_unmap_page (&proc->procgroup->pd, vpage, 0);
}
spin_unlock (&proc->pd->lock, &ctxprpd);
spin_unlock (&proc->procgroup->pd.lock, &ctxprpd);
return true;
}
@@ -155,10 +156,8 @@ struct elf_aux proc_load_segments (struct proc* proc, uint8_t* elf) {
size_t blks = div_align_up (phdr->p_memsz + off, PAGE_SIZE);
struct proc_resource_mem_init mem_init = {.pages = blks};
int rid = atomic_fetch_add (&proc->resources->sys_rids, 1);
struct proc_resource* r =
proc_create_resource (proc, rid, PR_MEM, RV_PRIVATE, (void*)&mem_init);
int rid = procgroup_get_sys_rid (proc->procgroup);
struct proc_resource* r = proc_create_resource_mem (proc->procgroup, rid, blks, 0, false);
if (r == NULL) {
DEBUG ("pmm oom error while loading ELF segments! (tried to alloc %zu blks)\n", blks);
}
@@ -357,14 +356,9 @@ static void proc_kpproc_init (void) {
kpproc.state = PROC_PSEUDO;
kpproc.pid = 0;
kpproc.resources = malloc (sizeof (*kpproc.resources));
kpproc.resources->tree = NULL;
kpproc.resources->lock = RW_SPIN_LOCK_INIT;
kpproc.resources->refs = 1;
kpproc.resources->sys_rids = 0;
kpproc.pd = mm_get_kernel_pd ();
kpproc.cpu = thiscpu;
kpproc.procgroup = procgroup_create ();
procgroup_attach (kpproc.procgroup, &kpproc);
rbtree_insert (struct proc, &proc_tree, &kpproc.proc_tree_link, proc_tree_link, pid);
@@ -394,9 +388,7 @@ static void proc_kpproc_init (void) {
memcpy ((struct kpproc_fb*)((uintptr_t)hhdm->offset + fb_info_memblk_paddr), &fb_info,
sizeof (fb_info));
struct proc_resource_mem_init mem_init = {
.pages = pages, .paddr = fb_info_memblk_paddr, .managed = true};
proc_create_resource (&kpproc, 0, PR_MEM, RV_PUBLIC, &mem_init);
proc_create_resource_mem (kpproc.procgroup, 0, pages, fb_info_memblk_paddr, true);
}
}

View File

@@ -6,6 +6,7 @@
#include <libk/list.h>
#include <libk/rbtree.h>
#include <libk/std.h>
#include <proc/procgroup.h>
#include <proc/resource.h>
#include <proc/suspension_q.h>
#include <sync/rw_spin_lock.h>
@@ -23,25 +24,11 @@
#define PROC_SUSPENDED 2
#define PROC_PSEUDO 3
/* process flags */
#define PROC_USTK_PREALLOC (1 << 0)
struct cpu;
struct proc_mapping {
struct list_node_link proc_mappings_link;
uintptr_t paddr;
uintptr_t vaddr;
size_t size;
};
struct proc_resources {
atomic_int refs;
atomic_int sys_rids;
struct rb_node_link* tree;
rw_spin_lock_t lock;
};
struct proc_sq_entry {
struct list_node_link sq_link;
struct list_node_link proc_link;
@@ -52,18 +39,16 @@ struct proc_sq_entry {
struct proc {
int pid;
struct rb_node_link proc_tree_link;
struct rb_node_link procgroup_memb_tree_link;
struct list_node_link cpu_run_q_link;
struct list_node_link reap_link;
struct list_node_link* sq_entries;
struct list_node_link* mappings; /* pd.lock implicitly protects this field */
struct procgroup* procgroup;
struct proc_platformdata pdata;
uint32_t flags;
struct pd* pd;
spin_lock_t lock;
struct cpu* cpu;
atomic_int state;
struct proc_resources* resources;
};
void proc_sched (void);

96
kernel/proc/procgroup.c Normal file
View File

@@ -0,0 +1,96 @@
#include <libk/rbtree.h>
#include <libk/std.h>
#include <mm/liballoc.h>
#include <proc/proc.h>
#include <proc/procgroup.h>
#include <sync/rw_spin_lock.h>
#include <sync/spin_lock.h>
#include <sys/debug.h>
#include <sys/mm.h>
static struct rb_node_link* procgroup_tree = NULL;
static rw_spin_lock_t procgroup_tree_lock = RW_SPIN_LOCK_INIT;
static atomic_int pgids = 0;
struct procgroup* procgroup_create (void) {
spin_lock_ctx_t ctxpgtr;
struct procgroup* procgroup = malloc (sizeof (*procgroup));
if (procgroup == NULL) {
return NULL;
}
procgroup->refs = 0;
procgroup->memb_proc_tree = NULL;
procgroup->lock = SPIN_LOCK_INIT;
procgroup->pgid = atomic_fetch_add (&pgids, 1);
procgroup->pd.lock = SPIN_LOCK_INIT;
procgroup->pd.cr3_paddr = mm_alloc_user_pd_phys ();
rw_spin_write_lock (&procgroup_tree_lock, &ctxpgtr);
rbtree_insert (struct procgroup, &procgroup_tree, &procgroup->procgroup_tree_link,
procgroup_tree_link, pgid);
rw_spin_write_unlock (&procgroup_tree_lock, &ctxpgtr);
return procgroup;
}
void procgroup_attach (struct procgroup* procgroup, struct proc* proc) {
spin_lock_ctx_t ctxpg, ctxpr;
spin_lock (&procgroup->lock, &ctxpg);
spin_lock (&proc->lock, &ctxpr);
rbtree_insert (struct proc, &procgroup->memb_proc_tree, &proc->procgroup_memb_tree_link,
procgroup_memb_tree_link, pid);
atomic_fetch_add (&procgroup->refs, 1);
DEBUG ("procgrpup attach PID %d to PGID %d\n", proc->pid, procgroup->pgid);
spin_unlock (&proc->lock, &ctxpr);
spin_unlock (&procgroup->lock, &ctxpg);
}
void procgroup_detach (struct procgroup* procgroup, struct proc* proc) {
spin_lock_ctx_t ctxpg, ctxpr, ctxpgtr;
spin_lock (&procgroup->lock, &ctxpg);
spin_lock (&proc->lock, &ctxpr);
rbtree_delete (&procgroup->memb_proc_tree, &proc->procgroup_memb_tree_link);
int refs = atomic_fetch_sub (&procgroup->refs, 1);
DEBUG ("procgrpup detach PID %d to PGID %d\n", proc->pid, procgroup->pgid);
spin_unlock (&proc->lock, &ctxpr);
spin_unlock (&procgroup->lock, &ctxpg);
if (refs == 1) {
rw_spin_write_lock (&procgroup_tree_lock, &ctxpgtr);
spin_lock (&procgroup->lock, &ctxpg);
rbtree_delete (&procgroup_tree, &procgroup->procgroup_tree_link);
spin_unlock (&procgroup->lock, &ctxpg);
rw_spin_write_unlock (&procgroup_tree_lock, &ctxpgtr);
/* unlink resources */
struct rb_node_link* rnode;
rbtree_first (&procgroup->resource_tree, rnode);
while (rnode) {
struct rb_node_link* next;
rbtree_next (rnode, next);
struct proc_resource* resource =
rbtree_entry (rnode, struct proc_resource, resource_tree_link);
rnode = next;
proc_resource_unlink (resource);
}
free (procgroup);
}
}
int procgroup_get_sys_rid (struct procgroup* procgroup) {
return atomic_fetch_add (&procgroup->sys_rids, 1);
}

38
kernel/proc/procgroup.h Normal file
View File

@@ -0,0 +1,38 @@
#ifndef _KERNEL_PROC_PROCGROUP_H
#define _KERNEL_PROC_PROCGROUP_H
#include <libk/list.h>
#include <libk/rbtree.h>
#include <libk/std.h>
#include <proc/resource.h>
#include <sync/spin_lock.h>
#include <sys/mm.h>
struct proc;
struct proc_mapping {
struct list_node_link proc_mappings_link;
uintptr_t paddr;
uintptr_t vaddr;
size_t size;
};
struct procgroup {
int pgid;
struct rb_node_link procgroup_tree_link;
struct rb_node_link* memb_proc_tree;
spin_lock_t lock;
atomic_int refs;
struct rb_node_link* resource_tree;
atomic_int sys_rids;
struct pd pd;
struct list_node_link* mappings; /* protected by pd.lock */
};
struct procgroup* procgroup_create (void);
void procgroup_attach (struct procgroup* procgroup, struct proc* proc);
void procgroup_detach (struct procgroup* procgroup, struct proc* proc);
int procgroup_get_sys_rid (struct procgroup* procgroup);
#endif // _KERNEL_PROC_PROCGROUP_H

View File

@@ -7,169 +7,98 @@
#include <mm/pmm.h>
#include <proc/mutex.h>
#include <proc/proc.h>
#include <proc/procgroup.h>
#include <proc/resource.h>
#include <sync/spin_lock.h>
#include <sys/debug.h>
static struct rb_node_link* resource_tree = NULL;
static rw_spin_lock_t resource_tree_lock = RW_SPIN_LOCK_INIT;
void proc_cleanup_resources (struct proc* proc) {
spin_lock_ctx_t ctxrs;
rw_spin_write_lock (&proc->resources->lock, &ctxrs);
struct rb_node_link* rnode;
rbtree_first (&proc->resources->tree, rnode);
while (rnode) {
struct rb_node_link* next;
rbtree_next (rnode, next);
struct proc_resource* resource =
rbtree_entry (rnode, struct proc_resource, local_resource_tree_link);
rnode = next;
proc_drop_resource (proc, resource, false);
}
rw_spin_write_unlock (&proc->resources->lock, &ctxrs);
if (atomic_fetch_sub (&proc->resources->refs, 1) == 1) {
free (proc->resources);
}
}
void proc_drop_resource (struct proc* proc, struct proc_resource* resource, bool lock) {
spin_lock_ctx_t ctxrs;
if (atomic_fetch_sub (&resource->refs, 1) == 1) {
DEBUG ("resource=%p created_by=%d vis=%d type=%d rid=%d refs=%d\n", resource,
resource->created_by_pid, resource->visibility, resource->type, resource->rid,
atomic_load (&resource->refs));
switch (resource->visibility) {
case RV_PRIVATE: {
if (lock)
rw_spin_write_lock (&proc->resources->lock, &ctxrs);
rbtree_delete (&proc->resources->tree, &resource->local_resource_tree_link);
if (lock)
rw_spin_write_unlock (&proc->resources->lock, &ctxrs);
} break;
case RV_PUBLIC: {
if (lock)
rw_spin_write_lock (&resource_tree_lock, &ctxrs);
rbtree_delete (&resource_tree, &resource->global_resource_tree_link);
if (lock)
rw_spin_write_unlock (&resource_tree_lock, &ctxrs);
} break;
default: {
assert (0);
} break;
}
resource->ops.cleanup (proc, resource);
free (resource);
}
}
struct proc_resource* proc_find_resource (struct proc* proc, int rid, int vis) {
struct proc_resource* proc_find_resource (struct procgroup* procgroup, int rid) {
spin_lock_ctx_t ctxpg;
struct proc_resource* resource = NULL;
spin_lock_ctx_t ctxrs;
switch (vis) {
case RV_PRIVATE: {
/* User wants to create a private resource, so search locally */
rw_spin_read_lock (&proc->resources->lock, &ctxrs);
rbtree_find (struct proc_resource, &proc->resources->tree, rid, resource,
local_resource_tree_link, rid);
rw_spin_read_unlock (&proc->resources->lock, &ctxrs);
} break;
case RV_PUBLIC: {
/* User wants to create a public resource, so search globally */
rw_spin_read_lock (&resource_tree_lock, &ctxrs);
rbtree_find (struct proc_resource, &resource_tree, rid, resource, global_resource_tree_link,
rid);
rw_spin_read_unlock (&resource_tree_lock, &ctxrs);
} break;
default: {
assert (0);
} break;
}
spin_lock (&procgroup->lock, &ctxpg);
rbtree_find (struct proc_resource, &procgroup->resource_tree, rid, resource, resource_tree_link,
rid);
spin_unlock (&procgroup->lock, &ctxpg);
return resource;
}
struct proc_resource* proc_create_resource (struct proc* proc, int rid, int type, int vis,
void* data) {
spin_lock_ctx_t ctxrs;
struct proc_resource* proc_create_resource_mem (struct procgroup* procgroup, int rid, size_t pages,
uintptr_t paddr, bool managed) {
spin_lock_ctx_t ctxpg;
struct proc_resource* resource;
/* Check if resource RID already exists */
struct proc_resource* resource_check = proc_find_resource (proc, rid, vis);
/* Resource was found either way, so it already exists */
if (resource_check != NULL)
if (pages == 0)
return NULL;
/* create the resource */
resource = proc_find_resource (procgroup, rid);
if (resource != NULL)
return resource;
struct proc_resource* resource = malloc (sizeof (*resource));
resource = malloc (sizeof (*resource));
if (resource == NULL)
return NULL;
memset (resource, 0, sizeof (*resource));
resource->lock = SPIN_LOCK_INIT;
resource->type = type;
resource->refs = 1;
resource->ops.cleanup = &proc_cleanup_resource_mem;
resource->rid = rid;
resource->visibility = vis;
resource->created_by_pid = proc->pid;
switch (resource->type) {
case PR_MEM: {
struct proc_resource_mem_init* mem_init = data;
proc_create_resource_mem (&resource->u.mem, mem_init);
resource->ops.cleanup = &proc_cleanup_resource_mem;
resource->u.mem.resource = resource;
DEBUG ("PR_MEM resource=%p created_by=%d, type=%d rid=%d paddr=%p, pages=%zu\n", resource,
resource->created_by_pid, resource->type, resource->rid, resource->u.mem.paddr,
resource->u.mem.pages);
} break;
case PR_MUTEX: {
proc_create_resource_mutex (&resource->u.mutex);
resource->ops.cleanup = &proc_cleanup_resource_mutex;
resource->u.mutex.resource = resource;
DEBUG ("PR_MUTEX resource=%p created_by=%d type=%d rid=%d\n", resource,
resource->created_by_pid, resource->type, resource->rid);
} break;
default: {
free (resource);
return NULL;
} break;
resource->type = PR_MEM;
resource->u.mem.resource = resource;
if (managed) {
resource->u.mem.managed = true;
} else {
paddr = pmm_alloc (pages);
if (paddr == PMM_ALLOC_ERR) {
free (resource);
return NULL;
}
resource->u.mem.managed = false;
}
resource->u.mem.paddr = paddr;
resource->u.mem.pages = resource->u.mem.alive_pages = pages;
resource->refs = 1;
switch (resource->visibility) {
case RV_PRIVATE: {
rw_spin_write_lock (&proc->resources->lock, &ctxrs);
rbtree_insert (struct proc_resource, &proc->resources->tree,
&resource->local_resource_tree_link, local_resource_tree_link, rid);
rw_spin_write_unlock (&proc->resources->lock, &ctxrs);
} break;
case RV_PUBLIC: {
rw_spin_write_lock (&resource_tree_lock, &ctxrs);
rbtree_insert (struct proc_resource, &resource_tree, &resource->global_resource_tree_link,
global_resource_tree_link, rid);
rw_spin_write_unlock (&resource_tree_lock, &ctxrs);
} break;
default: {
assert (0);
} break;
}
spin_lock (&procgroup->lock, &ctxpg);
rbtree_insert (struct proc_resource, &procgroup->resource_tree, &resource->resource_tree_link,
resource_tree_link, rid);
spin_unlock (&procgroup->lock, &ctxpg);
return resource;
}
struct proc_resource* proc_create_resource_mutex (struct procgroup* procgroup, int rid) {
spin_lock_ctx_t ctxpg;
struct proc_resource* resource;
resource = proc_find_resource (procgroup, rid);
if (resource != NULL)
return resource;
resource = malloc (sizeof (*resource));
if (resource == NULL)
return NULL;
memset (resource, 0, sizeof (*resource));
resource->lock = SPIN_LOCK_INIT;
resource->ops.cleanup = &proc_cleanup_resource_mutex;
resource->u.mem.resource = resource;
resource->rid = rid;
resource->type = PR_MUTEX;
resource->refs = 1;
spin_lock (&procgroup->lock, &ctxpg);
rbtree_insert (struct proc_resource, &procgroup->resource_tree, &resource->resource_tree_link,
resource_tree_link, rid);
spin_unlock (&procgroup->lock, &ctxpg);
return resource;
}
void proc_resource_unlink (struct proc_resource* resource) {
if (atomic_fetch_sub (&resource->refs, 1) == 1) {
resource->ops.cleanup (resource);
free (resource);
}
}

View File

@@ -11,33 +11,28 @@
#define PR_MEM 0
#define PR_MUTEX 1
#define RV_PRIVATE 0
#define RV_PUBLIC 1
struct proc;
struct procgroup;
struct proc_resource {
int type;
int rid;
int visibility;
spin_lock_t lock;
atomic_int refs;
struct rb_node_link global_resource_tree_link;
struct rb_node_link local_resource_tree_link;
spin_lock_t lock;
struct rb_node_link resource_tree_link;
union {
struct proc_resource_mem mem;
struct proc_mutex mutex;
} u;
struct {
void (*cleanup) (struct proc* proc, struct proc_resource* resource);
void (*cleanup) (struct proc_resource* resource);
} ops;
int created_by_pid;
};
struct proc_resource* proc_create_resource (struct proc* proc, int rid, int type, int vis,
void* data);
struct proc_resource* proc_find_resource (struct proc* proc, int rid, int vis);
void proc_drop_resource (struct proc* proc, struct proc_resource* resource, bool lock);
void proc_cleanup_resources (struct proc* proc);
struct proc_resource* proc_find_resource (struct procgroup* procgroup, int rid);
struct proc_resource* proc_create_resource_mem (struct procgroup* procgroup, int rid, size_t pages,
uintptr_t paddr, bool managed);
struct proc_resource* proc_create_resource_mutex (struct procgroup* procgroup, int rid);
void proc_resource_unlink (struct proc_resource* resource);
#endif // _KERNEL_PROC_RESOURCE_H

View File

@@ -1,9 +1,11 @@
c += proc/proc.c \
proc/resource.c \
proc/mutex.c \
proc/mem.c
proc/mem.c \
proc/procgroup.c
o += proc/proc.o \
proc/resource.o \
proc/mutex.o \
proc/mem.o
proc/mem.o \
proc/procgroup.o