Compare commits
15 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
e399ec6a46
|
|||
| dd9315f2f1 | |||
| f91831616c | |||
|
0240220796
|
|||
|
437bd0e751
|
|||
| 1fe5eb2d38 | |||
|
dbffb7f5fa
|
|||
|
aa30d9c6b5
|
|||
|
11a9dd4adb
|
|||
|
d644126901
|
|||
|
8be2a744b4
|
|||
|
e9b57f70b1
|
|||
|
e8a0a36889
|
|||
|
1fc5225fd2
|
|||
| 7d03a0090b |
@@ -16,11 +16,13 @@ endif
|
||||
OBJFILES := $(patsubst $(SRC)/%.c, $(BUILDDIR)/%.o, $(SOURCES))
|
||||
|
||||
CC := x86_64-elf-gcc
|
||||
CC_FLAGS=-Wall -Wextra -std=gnu99 -nostdlib -ffreestanding -fstack-protector -fno-omit-frame-pointer -fno-stack-check -fno-PIC -ffunction-sections -fdata-sections -mcmodel=kernel
|
||||
CC_FLAGS=-Wall -Wextra -std=gnu99 -nostdlib -ffreestanding -fstack-protector -fno-omit-frame-pointer -fno-stack-check -fno-PIC -ffunction-sections -fdata-sections -mcmodel=kernel -mno-red-zone
|
||||
|
||||
LD := x86_64-elf-ld
|
||||
|
||||
$(ELFFILE): $(BUILDDIR) $(OBJFILES)
|
||||
nasm -f bin user/hello.S -o $(BUILDDIR)/hello
|
||||
nasm -f bin user/pedicel.S -o $(BUILDDIR)/pedicel
|
||||
nasm -f elf64 src/arch/x86/idt.S -o $(BUILDDIR)/idt_stub.o
|
||||
$(LD) -o $(ELFFILE) -T linker.ld $(OBJFILES) $(BUILDDIR)/idt_stub.o
|
||||
# Get the symbols for debugging
|
||||
@@ -47,6 +49,8 @@ build-iso: limine/limine $(ELFFILE)
|
||||
cp -v $(ELFFILE) iso_root/boot
|
||||
mkdir -p iso_root/boot/limine
|
||||
cp -v limine.conf iso_root/boot/limine
|
||||
cp $(BUILDDIR)/hello iso_root/boot/
|
||||
cp $(BUILDDIR)/pedicel iso_root/boot/
|
||||
mkdir -p iso_root/EFI/BOOT
|
||||
cp -v limine/limine-bios.sys limine/limine-bios-cd.bin limine/limine-uefi-cd.bin iso_root/boot/limine/
|
||||
cp -v limine/BOOTX64.EFI iso_root/EFI/BOOT/
|
||||
|
||||
+9
-1
@@ -25,6 +25,9 @@ The recommended hardware to run PepperOS is the following:
|
||||
|
||||
## b. Features
|
||||
|
||||
- Round robin preemptive scheduling
|
||||
- Coexistence of ring 0 and ring 3 processes
|
||||
|
||||
## II. Kernel architecture
|
||||
|
||||
### a. Boot process
|
||||
@@ -37,4 +40,9 @@ The recommended hardware to run PepperOS is the following:
|
||||
|
||||
## III. Syscall table
|
||||
|
||||
Not yet implemented.
|
||||
The syscall interface in the Pepper kernel uses the System V ABI convention for argument order.
|
||||
|
||||
Name | Number (%rax) | arg0 (%rdi) | arg1 (%rsi) | arg2 (%rdx) |
|
||||
|---|---|---|---|---|
|
||||
| sys_write | 1 | unsigned int fd | const char* buf | size_t count | |
|
||||
| sys_exit | 60 | int error_code | | | |
|
||||
@@ -13,19 +13,32 @@
|
||||
// we'll only use this as a requirement for paging, not more.
|
||||
// This means base 0 and no limit (whole address space)
|
||||
|
||||
#define NUM_GDT_ENTRIES 5
|
||||
#define NUM_GDT_ENTRIES 7
|
||||
|
||||
#define NULL_SELECTOR 0x00
|
||||
#define KERNEL_CODE_SEGMENT 0x08
|
||||
#define KERNEL_DATA_SEGMENT 0x10
|
||||
#define USER_CODE_SEGMENT 0x18
|
||||
#define USER_DATA_SEGMENT 0x20
|
||||
#define TSS_SEGMENT 0x28
|
||||
|
||||
struct GDTR {
|
||||
uint16_t limit;
|
||||
uint64_t address;
|
||||
} __attribute__((packed));
|
||||
|
||||
struct tss {
|
||||
uint32_t reserved0;
|
||||
uint64_t rsp0;
|
||||
uint64_t rsp1;
|
||||
uint64_t rsp2;
|
||||
uint64_t reserved1;
|
||||
uint64_t ist[7];
|
||||
uint64_t reserved2;
|
||||
uint16_t reserved3;
|
||||
uint16_t iopb;
|
||||
} __attribute__((packed));
|
||||
|
||||
void gdt_init(void);
|
||||
|
||||
#endif
|
||||
+2
-3
@@ -10,7 +10,6 @@ void wrmsr(uint32_t msr, uint64_t value);
|
||||
bool x86_has_msr();
|
||||
void x86_arch_init();
|
||||
|
||||
|
||||
void x86_cpu_identification();
|
||||
int cpuid_get_vendor_string(char* str);
|
||||
|
||||
@@ -36,7 +35,7 @@ struct idtr {
|
||||
// All general-purpose registers (except rsp) as stored on the stack,
|
||||
// plus the values we pushed (vector number, error code) and the iret frame
|
||||
// In reverse order because the stack grows downwards.
|
||||
struct cpu_status_t {
|
||||
struct cpu_status {
|
||||
uint64_t r15;
|
||||
uint64_t r14;
|
||||
uint64_t r13;
|
||||
@@ -63,6 +62,6 @@ struct cpu_status_t {
|
||||
uint64_t iret_ss;
|
||||
};
|
||||
|
||||
struct cpu_status_t* syscall_handler(struct cpu_status_t* regs);
|
||||
struct cpu_status* syscall_handler(struct cpu_status* regs);
|
||||
|
||||
#endif
|
||||
+7
-2
@@ -9,8 +9,8 @@
|
||||
|
||||
/* version */
|
||||
#define PEPPEROS_VERSION_MAJOR "0"
|
||||
#define PEPPEROS_VERSION_MINOR "0"
|
||||
#define PEPPEROS_VERSION_PATCH "109"
|
||||
#define PEPPEROS_VERSION_MINOR "1"
|
||||
#define PEPPEROS_VERSION_PATCH "121"
|
||||
#define PEPPEROS_SPLASH \
|
||||
"\x1b[38;5;196m \x1b[38;5;231m____ _____\r\n\x1b[0m"\
|
||||
"\x1b[38;5;196m ____ ___ ____ ____ ___ _____\x1b[38;5;231m/ __ \\/ ___/\r\n\x1b[0m"\
|
||||
@@ -40,6 +40,11 @@
|
||||
#define KERNEL_STACK_SIZE 65536
|
||||
#define KERNEL_IDT_ENTRIES 33
|
||||
|
||||
/* user */
|
||||
#define USER_STACK_TOP 0x80000000
|
||||
#define USER_STACK_PAGES 16 // 16*4096 = 64kb
|
||||
#define USER_CODE_START 0x400000 // like linux
|
||||
|
||||
/* paging */
|
||||
#define PAGING_MAX_PHYS 0x200000000
|
||||
|
||||
|
||||
+2
-1
@@ -35,7 +35,7 @@ extern volatile uint64_t ticks;
|
||||
|
||||
// printf("debug: [%s]: " log "\n", __FILE__, ##__VA_ARGS__);
|
||||
|
||||
void panic(struct cpu_status_t* ctx, const char* str);
|
||||
void panic(struct cpu_status* ctx, const char* str);
|
||||
void hcf(void);
|
||||
void idle(void);
|
||||
|
||||
@@ -54,6 +54,7 @@ struct boot_context {
|
||||
struct limine_hhdm_response* hhdm;
|
||||
struct limine_kernel_address_response* kaddr;
|
||||
struct limine_boot_time_response* bootdate;
|
||||
struct limine_module_response* module;
|
||||
};
|
||||
|
||||
// Are these modules initialized yet?
|
||||
|
||||
+3
-2
@@ -16,11 +16,11 @@
|
||||
#include <stddef.h>
|
||||
#include <stdint.h>
|
||||
|
||||
struct heap_block_t {
|
||||
struct heap_block {
|
||||
size_t size;
|
||||
bool free; // 1byte
|
||||
uint8_t reserved[7]; // (7+1 = 8 bytes)
|
||||
struct heap_block_t* next;
|
||||
struct heap_block* next;
|
||||
} __attribute__((aligned(16)));
|
||||
|
||||
void kheap_init(void);
|
||||
@@ -28,5 +28,6 @@ void* kmalloc(size_t size);
|
||||
void kfree(void* ptr);
|
||||
void* kalloc_stack(void);
|
||||
void kheap_map_page(void);
|
||||
void kheap_info();
|
||||
|
||||
#endif
|
||||
@@ -34,7 +34,7 @@ extern uint64_t hhdm_off;
|
||||
#define PAGE_ALIGN_DOWN(x) ((x) & PTE_ADDR_MASK)
|
||||
|
||||
#define ALIGN(size) ALIGN_UP(size, 16)
|
||||
#define BLOCK_MIN_SIZE (sizeof(struct heap_block_t) + 16)
|
||||
#define BLOCK_MIN_SIZE (sizeof(struct heap_block) + 16)
|
||||
|
||||
#define PML4_INDEX(x) (((x) >> 39) & 0x1FF)
|
||||
#define PDPT_INDEX(x) (((x) >> 30) & 0x1FF)
|
||||
|
||||
@@ -22,6 +22,8 @@ void vmm_unmap(uint64_t* pml4, uint64_t virt);
|
||||
void* vmm_map(uint64_t* pml4, uint64_t virt, uint64_t flags);
|
||||
uint64_t* vmm_create_address_space();
|
||||
uint64_t vmm_virt_to_phys(uint64_t* pml4, uint64_t virt);
|
||||
uintptr_t vmm_alloc_user_stack(uint64_t* pml4);
|
||||
uintptr_t vmm_alloc_user_code(uint64_t* pml4, void* code_addr, uint64_t code_size);
|
||||
|
||||
#define VMM_USER_SPACE_START 0x0000000000001000
|
||||
#define VMM_USER_SPACE_END 0x00007FFFFFFFF000
|
||||
|
||||
+12
-8
@@ -10,6 +10,7 @@
|
||||
#include <stddef.h>
|
||||
#include <config.h>
|
||||
#include <stdint.h>
|
||||
#include <limine.h>
|
||||
|
||||
typedef enum {
|
||||
READY,
|
||||
@@ -17,23 +18,26 @@ typedef enum {
|
||||
DEAD
|
||||
} status_t;
|
||||
|
||||
struct process_t {
|
||||
struct process {
|
||||
size_t pid;
|
||||
char name[PROCESS_NAME_MAX];
|
||||
|
||||
status_t status;
|
||||
struct cpu_status_t* context;
|
||||
struct cpu_status* context;
|
||||
void* root_page_table; // Process PML4 (should contain kernel PML4 in higher half [256-511]
|
||||
struct process_t* next;
|
||||
void* kernel_stack; // Used for interrupts (syscall: int 0x80), defines the TSS RSP0
|
||||
struct process* next;
|
||||
};
|
||||
|
||||
void process_init(void);
|
||||
struct process_t* process_create(char* name, void(*function)(void*), void* arg);
|
||||
void process_add(struct process_t** processes_list, struct process_t* process);
|
||||
void process_delete(struct process_t** processes_list, struct process_t* process);
|
||||
struct process_t* process_get_next(struct process_t* process);
|
||||
struct process* process_create(char* name, void(*function)(void*), void* arg);
|
||||
void process_add(struct process** processes_list, struct process* process);
|
||||
void process_delete(struct process** processes_list, struct process* process);
|
||||
struct process* process_get_next(struct process* process);
|
||||
void process_exit(void);
|
||||
|
||||
void process_display_list(struct process_t* processes_list);
|
||||
void process_display_list(struct process* processes_list);
|
||||
|
||||
void process_create_user(struct limine_file* file, char* name);
|
||||
|
||||
#endif
|
||||
|
||||
@@ -7,7 +7,7 @@
|
||||
#ifndef SCHEDULER_H
|
||||
#define SCHEDULER_H
|
||||
|
||||
struct cpu_status_t* scheduler_schedule(struct cpu_status_t* context);
|
||||
struct cpu_status* scheduler_schedule(struct cpu_status* context);
|
||||
void scheduler_init(void);
|
||||
|
||||
#endif
|
||||
@@ -10,13 +10,13 @@
|
||||
#include <stdbool.h>
|
||||
#include <stdint.h>
|
||||
|
||||
struct spinlock_t
|
||||
struct spinlock
|
||||
{
|
||||
bool locked;
|
||||
uint64_t rflags;
|
||||
};
|
||||
|
||||
void spinlock_acquire(struct spinlock_t* lock);
|
||||
void spinlock_release(struct spinlock_t* lock);
|
||||
void spinlock_acquire(struct spinlock* lock);
|
||||
void spinlock_release(struct spinlock* lock);
|
||||
|
||||
#endif
|
||||
@@ -62,5 +62,4 @@ struct ubsan_overflow_data
|
||||
struct ubsan_type_descriptor* type;
|
||||
};
|
||||
|
||||
|
||||
#endif
|
||||
@@ -6,3 +6,5 @@ interface_branding: Welcome to the PepperOS disk!
|
||||
|
||||
comment: Default configuration (warning: spicy)
|
||||
path: boot():/boot/pepperk
|
||||
module_path: boot():/boot/hello
|
||||
module_path: boot():/boot/pedicel
|
||||
@@ -0,0 +1,138 @@
|
||||
/*
|
||||
* @author xamidev <xamidev@riseup.net>
|
||||
* @brief Global Descriptor Table and Task State Segment setup
|
||||
* @license GPL-3.0-only
|
||||
*/
|
||||
|
||||
#include <arch/gdt.h>
|
||||
#include <stdint.h>
|
||||
#include <io/serial/serial.h>
|
||||
#include <kernel.h>
|
||||
|
||||
// Descriptors are 8-byte wide (64bits)
|
||||
// So the selectors will be (in bytes): 0x0, 0x8, 0x10, 0x18, etc..
|
||||
uint64_t gdt_entries[NUM_GDT_ENTRIES];
|
||||
struct GDTR gdtr;
|
||||
|
||||
struct tss tss = {0};
|
||||
|
||||
/*
|
||||
* gdt_load - Loads Global Descriptor Table
|
||||
*/
|
||||
static void gdt_load()
|
||||
{
|
||||
asm("lgdt %0" : : "m"(gdtr));
|
||||
}
|
||||
|
||||
/*
|
||||
* gdt_flush - Flushes the Global Descriptor Table
|
||||
*
|
||||
* This function loads new Segment Selectors to make
|
||||
* the GDT changes take effect
|
||||
*/
|
||||
static void gdt_flush()
|
||||
{
|
||||
// Here, 0x8 is the kernel code selector
|
||||
// and 0x10 is the kernel data selector
|
||||
asm volatile (
|
||||
"mov $0x10, %%ax \n" // Reload segments with kernel data selector
|
||||
"mov %%ax, %%ds \n"
|
||||
"mov %%ax, %%es \n"
|
||||
"mov %%ax, %%fs \n"
|
||||
"mov %%ax, %%gs \n"
|
||||
"mov %%ax, %%ss \n"
|
||||
|
||||
"pushq $0x8 \n" // CS reload
|
||||
"lea 1f(%%rip), %%rax \n"
|
||||
"push %%rax \n"
|
||||
"lretq \n"
|
||||
"1: \n" // Execution continues here after CS reload
|
||||
:
|
||||
:
|
||||
: "rax", "memory"
|
||||
);
|
||||
}
|
||||
|
||||
/*
|
||||
* get_set_entry - Sets a GDT entry
|
||||
* @num: Number of the entry (index in GDT)
|
||||
* @flags: Flags (Granularity, Size, Long mode)
|
||||
* @access: Access byte (contains Descriptor Privilege Level)
|
||||
*
|
||||
* This function fills a GDT entry with the specified @flags
|
||||
* and @access byte. The base and limit fields are left to zero
|
||||
* because we don't use segmentation for memory management.
|
||||
*/
|
||||
|
||||
static void gdt_set_entry(int num, uint8_t flags, uint8_t access)
|
||||
{
|
||||
uint64_t gdt_entry = 0;
|
||||
|
||||
gdt_entry |= (access << 8);
|
||||
gdt_entry |= (flags << 20);
|
||||
|
||||
// Rest (base, limit) is always zero
|
||||
gdt_entries[num] = gdt_entry << 32;
|
||||
}
|
||||
|
||||
/*
|
||||
* gdt_set_tss - Setup the TSS entry in the GDT
|
||||
* @num: Number of the entry (index in GDT)
|
||||
*
|
||||
* This function sets up a Task State Segment entry
|
||||
* in the Global Descriptor Table.
|
||||
*
|
||||
* The entry is 128-bit long, so it actually takes
|
||||
* two 64-bit GDT entries.
|
||||
*/
|
||||
static void gdt_set_tss(int num)
|
||||
{
|
||||
uint64_t tss_base = (uint64_t)&tss;
|
||||
uint64_t tss_limit = sizeof(struct tss) - 1;
|
||||
|
||||
tss.iopb = sizeof(struct tss);
|
||||
|
||||
uint64_t tss_low = 0;
|
||||
tss_low |= (tss_limit & 0xFFFFULL);
|
||||
tss_low |= (tss_base & 0xFFFFFFULL) << 16;
|
||||
tss_low |= 0x89ULL << 40;
|
||||
tss_low |= ((tss_limit >> 16) & 0xFULL) << 48;
|
||||
tss_low |= ((tss_base >> 24) & 0xFFULL) << 56;
|
||||
|
||||
uint64_t tss_high = (tss_base >> 32) & 0xFFFFFFFFULL;
|
||||
|
||||
gdt_entries[num] = tss_low;
|
||||
gdt_entries[num + 1] = tss_high;
|
||||
}
|
||||
|
||||
/*
|
||||
* gdt_init - Global Descriptor Table initialization
|
||||
*
|
||||
* This function loads a new GDT in the CPU.
|
||||
* It contains a null descriptor, kernel code and data
|
||||
* segments, and user code and data segments.
|
||||
* However, we do not use segmentation to manage memory on
|
||||
* 64-bit x86, as it's deprecated. Instead, we use paging.
|
||||
*/
|
||||
void gdt_init()
|
||||
{
|
||||
gdt_set_entry(0, 0, 0); // Null descriptor (0x0)
|
||||
gdt_set_entry(1, 0xA, 0x9B); // Kernel code (0x8)
|
||||
gdt_set_entry(2, 0xC, 0x93); // Kernel data (0x10)
|
||||
gdt_set_entry(3, 0xA, 0xFB); // User code (0x18)
|
||||
gdt_set_entry(4, 0xC, 0xF3); // User data (0x20)
|
||||
gdt_set_tss(5); // TSS (0x28)
|
||||
|
||||
// The -1 subtraction is some wizardry explained in the OSDev wiki -> GDT
|
||||
gdtr.limit = NUM_GDT_ENTRIES * sizeof(uint64_t) - 1;
|
||||
gdtr.address = (uint64_t)gdt_entries;
|
||||
|
||||
// Load the GDT we created, flush the old one
|
||||
gdt_load();
|
||||
gdt_flush();
|
||||
|
||||
// Load task register with new TSS
|
||||
asm volatile("ltr %%ax" : : "a"(TSS_SEGMENT) : "memory");
|
||||
|
||||
DEBUG("GDT initialized");
|
||||
}
|
||||
+4
-4
@@ -77,7 +77,7 @@ void idt_init()
|
||||
idt_set_entry(i, vector_0_handler + (i*16), 0);
|
||||
}
|
||||
|
||||
idt_set_entry(0x80, vector_128_handler, 0);
|
||||
idt_set_entry(0x80, vector_128_handler, 3);
|
||||
|
||||
idt_load(&idt);
|
||||
DEBUG("IDT initialized");
|
||||
@@ -108,7 +108,7 @@ static inline uint64_t read_cr2(void)
|
||||
* Also displays an interpretation of the thrown error code.
|
||||
* Then halts the system. We could implement demand paging later.
|
||||
*/
|
||||
static void page_fault_handler(struct cpu_status_t* ctx)
|
||||
static void page_fault_handler(struct cpu_status* ctx)
|
||||
{
|
||||
// It could be used to remap pages etc. to fix the fault, but right now what I'm more
|
||||
// interested in is getting more info out of those numbers cause i'm lost each time i have
|
||||
@@ -149,7 +149,7 @@ static void page_fault_handler(struct cpu_status_t* ctx)
|
||||
* Shows detail about a General Protection Fault,
|
||||
* and what may have caused it. Halts the system.
|
||||
*/
|
||||
static void gp_fault_handler(struct cpu_status_t* ctx)
|
||||
static void gp_fault_handler(struct cpu_status* ctx)
|
||||
{
|
||||
DEBUG("\x1b[38;5;231mGeneral Protection Fault at rip=0x%p, err=%u (%s)\x1b[0m",
|
||||
ctx->iret_rip,
|
||||
@@ -185,7 +185,7 @@ static void gp_fault_handler(struct cpu_status_t* ctx)
|
||||
* Return:
|
||||
* <context> - CPU context after interrupt
|
||||
*/
|
||||
struct cpu_status_t* interrupt_dispatch(struct cpu_status_t* context)
|
||||
struct cpu_status* interrupt_dispatch(struct cpu_status* context)
|
||||
{
|
||||
if (context == NULL) {
|
||||
panic(NULL, "Interrupt dispatch recieved NULL context!");
|
||||
|
||||
+22
-1
@@ -4,7 +4,7 @@
|
||||
* @license GPL-3.0-only
|
||||
*/
|
||||
|
||||
#include <mem/gdt.h>
|
||||
#include <arch/gdt.h>
|
||||
#include <stdint.h>
|
||||
#include <arch/x86.h>
|
||||
#include <kernel.h>
|
||||
@@ -29,6 +29,26 @@ static void x86_overwrite_pat()
|
||||
wrmsr(0x277, pat);
|
||||
}
|
||||
|
||||
/*
|
||||
* x86_enable_fpu - Enable Floating Point Unit
|
||||
*
|
||||
* This function enables the Floating Point Unit,
|
||||
* which allows the CPU to do floating point
|
||||
* operations.
|
||||
*
|
||||
* Here we do not check for FPU support but we
|
||||
* should. However most processors support it.
|
||||
*/
|
||||
static void x86_enable_fpu()
|
||||
{
|
||||
size_t cr4;
|
||||
__asm__ volatile("mov %%cr4, %0" : "=r"(cr4));
|
||||
cr4 |= 0x200;
|
||||
__asm__ volatile("mov %0, %%cr4" :: "r"(cr4));
|
||||
uint16_t cw = 0x37F; // control word
|
||||
asm volatile("fldcw %0" :: "m"(cw));
|
||||
}
|
||||
|
||||
/*
|
||||
* x86_arch_init - Initialize x86 CPU structures
|
||||
*
|
||||
@@ -42,6 +62,7 @@ static void x86_overwrite_pat()
|
||||
void x86_arch_init()
|
||||
{
|
||||
x86_overwrite_pat();
|
||||
x86_enable_fpu();
|
||||
x86_cpu_identification();
|
||||
idt_init();
|
||||
gdt_init();
|
||||
|
||||
+58
-7
@@ -4,21 +4,72 @@
|
||||
* @license GPL-3.0-only
|
||||
*/
|
||||
|
||||
#include "sched/scheduler.h"
|
||||
#include <arch/x86.h>
|
||||
#include <kernel.h>
|
||||
#include <stddef.h>
|
||||
#include <io/term/term.h>
|
||||
#include <sched/process.h>
|
||||
|
||||
struct cpu_status_t* syscall_handler(struct cpu_status_t* regs)
|
||||
{
|
||||
DEBUG("Syscall %lx with argument %lx", regs->rdi, regs->rsi);
|
||||
extern struct process* current_process;
|
||||
|
||||
switch (regs->rdi)
|
||||
void sys_write(unsigned int fd, const char* buf, size_t count)
|
||||
{
|
||||
case 0:
|
||||
switch (fd) {
|
||||
case 1: //stdout
|
||||
for (size_t i=0; i<count; i++) {
|
||||
internal_putc(buf[i], NULL);
|
||||
}
|
||||
break;
|
||||
case 1:
|
||||
|
||||
case 2: //stderr
|
||||
for (size_t i=0; i<count; i++) {
|
||||
internal_putc(buf[i], NULL);
|
||||
}
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
void sys_exit(int error_code)
|
||||
{
|
||||
current_process->status = DEAD;
|
||||
DEBUG("exiting process PID=%u name=%s", current_process->pid, current_process->name);
|
||||
}
|
||||
|
||||
/*
|
||||
* syscall_handler - System call dispatcher
|
||||
* @regs: CPU state
|
||||
*
|
||||
* This function is called from the interrupt dispatcher,
|
||||
* when an interrupt 0x80 is emitted from userland.
|
||||
*
|
||||
* It switches control to the syscall number provided
|
||||
* in %rax.
|
||||
*
|
||||
* We try to follow the System V convention here:
|
||||
* - syscall number in %rax
|
||||
* - args in %rdi, %rsi, %rdx, %r10, %r8, %r9
|
||||
* - return value (if any) in %rax
|
||||
*
|
||||
* Return:
|
||||
* <regs> - CPU state after system call
|
||||
*/
|
||||
struct cpu_status* syscall_handler(struct cpu_status* regs)
|
||||
{
|
||||
DEBUG("Syscall %lx with (arg0=%lx arg1=%lx)", regs->rax, regs->rdi, regs->rsi);
|
||||
|
||||
switch (regs->rax)
|
||||
{
|
||||
case 0: //sys_read
|
||||
break;
|
||||
case 1: //sys_write
|
||||
sys_write(regs->rdi, (char*)regs->rsi, regs->rdx);
|
||||
break;
|
||||
case 60: //sys_exit
|
||||
sys_exit(regs->rdi);
|
||||
break;
|
||||
default:
|
||||
regs->rsi = 0xdeadbeef;
|
||||
regs->rax = 0xbad515ca11;
|
||||
break;
|
||||
}
|
||||
|
||||
|
||||
+9
-2
@@ -13,6 +13,9 @@
|
||||
#include <limine.h>
|
||||
#include <stddef.h>
|
||||
|
||||
__attribute__((used, section(".limine_requests_start")))
|
||||
volatile LIMINE_REQUESTS_START_MARKER;
|
||||
|
||||
__attribute__((used, section(".limine_requests")))
|
||||
volatile struct limine_framebuffer_request framebuffer_request = {
|
||||
.id = LIMINE_FRAMEBUFFER_REQUEST,
|
||||
@@ -43,8 +46,11 @@ volatile struct limine_boot_time_request date_request = {
|
||||
.revision = 0
|
||||
};
|
||||
|
||||
__attribute__((used, section(".limine_requests_start")))
|
||||
volatile LIMINE_REQUESTS_START_MARKER;
|
||||
__attribute__((used, section(".limine_requests")))
|
||||
volatile struct limine_module_request module_request = {
|
||||
.id = LIMINE_MODULE_REQUEST,
|
||||
.revision = 0
|
||||
};
|
||||
|
||||
__attribute__((used, section(".limine_requests_end")))
|
||||
volatile LIMINE_REQUESTS_END_MARKER;
|
||||
@@ -58,4 +64,5 @@ void populate_boot_context(struct boot_context* ctx)
|
||||
ctx->hhdm = hhdm_request.response ? hhdm_request.response : NULL;
|
||||
ctx->kaddr = kerneladdr_request.response ? kerneladdr_request.response : NULL;
|
||||
ctx->bootdate = date_request.response ? date_request.response : NULL;
|
||||
ctx->module = module_request.response ? module_request.response : NULL;
|
||||
}
|
||||
+2
-2
@@ -18,7 +18,7 @@ extern int panic_count;
|
||||
*/
|
||||
void read_rflags(uint64_t rflags)
|
||||
{
|
||||
DEBUG("\x1b[38;5;226m%s%s%s%s%s%s%s%s%s%s%s%s%s%s%s%s%s\x1b[38;5;231m",
|
||||
printf("\x1b[38;5;226m%s%s%s%s%s%s%s%s%s%s%s%s%s%s%s%s%s\x1b[38;5;231m",
|
||||
CHECK_BIT(rflags, 0) ? "CF " : "", /*carry flag*/
|
||||
CHECK_BIT(rflags, 2) ? "PF " : "", /*parity flag*/
|
||||
CHECK_BIT(rflags, 4) ? "AF " : "", /*auxiliary carry flag*/
|
||||
@@ -47,7 +47,7 @@ void read_rflags(uint64_t rflags)
|
||||
* Will display to terminal if it is initialized, otherwise serial only.
|
||||
* Can be called with or without a CPU context.
|
||||
*/
|
||||
void panic(struct cpu_status_t* ctx, const char* str)
|
||||
void panic(struct cpu_status* ctx, const char* str)
|
||||
{
|
||||
CLEAR_INTERRUPTS;
|
||||
panic_count += 1;
|
||||
|
||||
+13
-7
@@ -289,15 +289,21 @@ int keyboard_getline(char* output, size_t size)
|
||||
|
||||
// Read until Enter is pressed
|
||||
while ((c = keyboard_getchar()) != 0x0A) {
|
||||
if (index == size-1) {
|
||||
output[index] = c;
|
||||
output[index+1] = '\0';
|
||||
return index;
|
||||
if (c == '\b') {
|
||||
if (index > 0) {
|
||||
index--;
|
||||
output[index] = '\0';
|
||||
printf(" \b");
|
||||
}
|
||||
output[index] = c;
|
||||
index++;
|
||||
continue;
|
||||
}
|
||||
output[index+1] = '\0';
|
||||
|
||||
if (index >= size-1) {
|
||||
continue;
|
||||
}
|
||||
output[index++] = c;
|
||||
}
|
||||
output[index] = '\0';
|
||||
return index;
|
||||
}
|
||||
|
||||
|
||||
@@ -11,7 +11,7 @@
|
||||
extern struct init_status init;
|
||||
|
||||
extern int panic_count;
|
||||
struct spinlock_t serial_lock = {0};
|
||||
struct spinlock serial_lock = {0};
|
||||
|
||||
/*
|
||||
* outb - Writes a byte to a CPU port
|
||||
|
||||
+2
-2
@@ -29,8 +29,8 @@ because this shitty implementation will be replaced one day by Flanterm
|
||||
extern struct flanterm_context* ft_ctx;
|
||||
extern struct init_status init;
|
||||
|
||||
struct spinlock_t term_lock = {0};
|
||||
struct spinlock_t printf_lock = {0};
|
||||
struct spinlock term_lock = {0};
|
||||
struct spinlock printf_lock = {0};
|
||||
|
||||
extern int panic_count;
|
||||
|
||||
|
||||
+8
-1
@@ -11,6 +11,7 @@
|
||||
#include <stdint.h>
|
||||
#include <kernel.h>
|
||||
#include <time/date.h>
|
||||
#include <mem/kheap.h>
|
||||
|
||||
__attribute__((noinline))
|
||||
void smash_it()
|
||||
@@ -49,7 +50,8 @@ void pedicel_main(void* arg)
|
||||
"syscall - trigger int 0x80\r\n"
|
||||
"pf - trigger a page fault\r\n"
|
||||
"now - get current date\r\n"
|
||||
"smash - smash the stack\r\n");
|
||||
"smash - smash the stack\r\n"
|
||||
"mem - get used heap info\r\n");
|
||||
continue;
|
||||
}
|
||||
|
||||
@@ -89,6 +91,11 @@ void pedicel_main(void* arg)
|
||||
continue;
|
||||
}
|
||||
|
||||
if (strncmp(input_buf, "mem", 3) == 0) {
|
||||
kheap_info();
|
||||
continue;
|
||||
}
|
||||
|
||||
printf("%s: command not found\r\n", input_buf);
|
||||
}
|
||||
}
|
||||
+19
-5
@@ -9,7 +9,7 @@
|
||||
#include <limine.h>
|
||||
#include <io/term/term.h>
|
||||
#include <io/serial/serial.h>
|
||||
#include <mem/gdt.h>
|
||||
#include <arch/gdt.h>
|
||||
#include <mem/utils.h>
|
||||
#include <kernel.h>
|
||||
#include <time/timer.h>
|
||||
@@ -62,10 +62,13 @@ extern volatile struct limine_memmap_request memmap_request;
|
||||
extern volatile struct limine_hhdm_request hhdm_request;
|
||||
extern volatile struct limine_kernel_address_request kerneladdr_request;
|
||||
extern volatile struct limine_boot_time_request date_request;
|
||||
extern volatile struct limine_module_request module_request;
|
||||
|
||||
extern struct process_t* processes_list;
|
||||
extern struct process_t* current_process;
|
||||
struct process_t* idle_proc;
|
||||
struct limine_file* file;
|
||||
|
||||
extern struct process* processes_list;
|
||||
extern struct process* current_process;
|
||||
struct process* idle_proc;
|
||||
|
||||
void idle_main(void* arg)
|
||||
{
|
||||
@@ -115,8 +118,19 @@ void kmain()
|
||||
|
||||
process_init();
|
||||
idle_proc = process_create("idle", (void*)idle_main, 0);
|
||||
process_create("pedicel", (void*)pedicel_main, 0);
|
||||
|
||||
if (!boot_ctx.module) {
|
||||
panic(NULL, "could not load 'hello' executable :(");
|
||||
}
|
||||
if (boot_ctx.module->module_count == 2) {
|
||||
file = boot_ctx.module->modules[0];
|
||||
DEBUG("file: addr=%p size=%u", file->address, file->size);
|
||||
process_create_user(file, "hello");
|
||||
|
||||
file = boot_ctx.module->modules[1];
|
||||
process_create_user(file, "pedicel");
|
||||
}
|
||||
process_create("kshell", (void*)pedicel_main, 0);
|
||||
scheduler_init();
|
||||
|
||||
printf(PEPPEROS_SPLASH);
|
||||
|
||||
-107
@@ -1,107 +0,0 @@
|
||||
/*
|
||||
* @author xamidev <xamidev@riseup.net>
|
||||
* @brief Global Descriptor Table (for legacy reasons)
|
||||
* @license GPL-3.0-only
|
||||
*/
|
||||
|
||||
#include <mem/gdt.h>
|
||||
#include <stdint.h>
|
||||
#include <io/serial/serial.h>
|
||||
#include <kernel.h>
|
||||
|
||||
// Descriptors are 8-byte wide (64bits)
|
||||
// So the selectors will be (in bytes): 0x0, 0x8, 0x10, 0x18, etc..
|
||||
uint64_t gdt_entries[NUM_GDT_ENTRIES];
|
||||
struct GDTR gdtr;
|
||||
|
||||
/*
|
||||
* gdt_load - Loads Global Descriptor Table
|
||||
*/
|
||||
static void gdt_load()
|
||||
{
|
||||
asm("lgdt %0" : : "m"(gdtr));
|
||||
}
|
||||
|
||||
/*
|
||||
* gdt_flush - Flushes the Global Descriptor Table
|
||||
*
|
||||
* This function loads new Segment Selectors to make
|
||||
* the GDT changes take effect
|
||||
*/
|
||||
static void gdt_flush()
|
||||
{
|
||||
// Here, 0x8 is the kernel code selector
|
||||
// and 0x10 is the kernel data selector
|
||||
asm volatile (
|
||||
"mov $0x10, %%ax \n" // Reload segments with kernel data selector
|
||||
"mov %%ax, %%ds \n"
|
||||
"mov %%ax, %%es \n"
|
||||
"mov %%ax, %%fs \n"
|
||||
"mov %%ax, %%gs \n"
|
||||
"mov %%ax, %%ss \n"
|
||||
|
||||
"pushq $0x8 \n" // CS reload
|
||||
"lea 1f(%%rip), %%rax \n"
|
||||
"push %%rax \n"
|
||||
"lretq \n"
|
||||
"1: \n" // Execution continues here after CS reload
|
||||
:
|
||||
:
|
||||
: "rax", "memory"
|
||||
);
|
||||
}
|
||||
|
||||
/*
|
||||
* gdt_init - Global Descriptor Table initialization
|
||||
*
|
||||
* This function loads a new GDT in the CPU.
|
||||
* It contains a null descriptor, kernel code and data
|
||||
* segments, and user code and data segments.
|
||||
* However, we do not use segmentation to manage memory on
|
||||
* 64-bit x86, as it's deprecated. Instead, we use paging.
|
||||
*/
|
||||
void gdt_init()
|
||||
{
|
||||
// Null descriptor (required)
|
||||
gdt_entries[0] = 0;
|
||||
|
||||
// Kernel code segment
|
||||
uint64_t kernel_code = 0;
|
||||
kernel_code |= 0b1101 << 8; // Selector type: accessed, read-enable, no conforming
|
||||
kernel_code |= 1 << 12; // not a system descriptor
|
||||
kernel_code |= 0 << 13; // DPL field = 0
|
||||
kernel_code |= 1 << 15; // Present
|
||||
kernel_code |= 1 << 21; // Long mode
|
||||
|
||||
// Left shift 32 bits so we place our stuff in the upper 32 bits of the descriptor.
|
||||
// The lower 32 bits contain limit and part of base and therefore are ignored in Long Mode
|
||||
// (because we'll use paging; segmentation is used only for legacy)
|
||||
gdt_entries[1] = kernel_code << 32;
|
||||
|
||||
uint64_t kernel_data = 0;
|
||||
kernel_data |= 0b0011 << 8;
|
||||
kernel_data |= 1 << 12;
|
||||
kernel_data |= 0 << 13;
|
||||
kernel_data |= 1 << 15;
|
||||
kernel_data |= 1 << 21;
|
||||
|
||||
gdt_entries[2] = kernel_data << 32;
|
||||
|
||||
// We re-use the kernel descriptors here, and just update their DPL fields
|
||||
// (Descriptor privilege level) from ring 0 -> to ring 3 (userspace)
|
||||
uint64_t user_code = kernel_code | (3 << 13);
|
||||
gdt_entries[3] = user_code;
|
||||
|
||||
uint64_t user_data = kernel_data | (3 << 13);
|
||||
gdt_entries[4] = user_data;
|
||||
|
||||
// The -1 subtraction is some wizardry explained in the OSDev wiki -> GDT
|
||||
gdtr.limit = NUM_GDT_ENTRIES * sizeof(uint64_t) - 1;
|
||||
gdtr.address = (uint64_t)gdt_entries;
|
||||
|
||||
// Load the GDT we created, flush the old one
|
||||
gdt_load();
|
||||
gdt_flush();
|
||||
|
||||
DEBUG("GDT initialized");
|
||||
}
|
||||
+39
-10
@@ -17,7 +17,7 @@ extern uint64_t kernel_virt_base;
|
||||
|
||||
uintptr_t kheap_start;
|
||||
|
||||
static struct heap_block_t* head = NULL;
|
||||
static struct heap_block* head = NULL;
|
||||
static uintptr_t end;
|
||||
|
||||
// Kernel root table (level 4)
|
||||
@@ -55,8 +55,8 @@ void kheap_init()
|
||||
end = current_addr;
|
||||
|
||||
// Give linked list head its properties
|
||||
head = (struct heap_block_t*)kheap_start;
|
||||
head->size = (end-kheap_start) - sizeof(struct heap_block_t);
|
||||
head = (struct heap_block*)kheap_start;
|
||||
head->size = (end-kheap_start) - sizeof(struct heap_block);
|
||||
head->free = true;
|
||||
head->next = NULL;
|
||||
DEBUG("Kernel heap initialized, head=0x%p, size=%u bytes", head, head->size);
|
||||
@@ -80,16 +80,16 @@ void* kmalloc(size_t size)
|
||||
if (!size) return NULL;
|
||||
size = ALIGN(size);
|
||||
|
||||
struct heap_block_t* curr = head;
|
||||
struct heap_block* curr = head;
|
||||
|
||||
while (curr) {
|
||||
// Is block free and big enough for us?
|
||||
if (curr->free && curr->size >= size) {
|
||||
// We split the block if it is big enough
|
||||
if (curr->size >= size + sizeof(struct heap_block_t) + 16) {
|
||||
struct heap_block_t* split = (struct heap_block_t*)((uintptr_t)curr + sizeof(struct heap_block_t) + size);
|
||||
if (curr->size >= size + sizeof(struct heap_block) + 16) {
|
||||
struct heap_block* split = (struct heap_block*)((uintptr_t)curr + sizeof(struct heap_block) + size);
|
||||
|
||||
split->size = curr->size - size - sizeof(struct heap_block_t);
|
||||
split->size = curr->size - size - sizeof(struct heap_block);
|
||||
split->free = true;
|
||||
split->next = curr->next;
|
||||
|
||||
@@ -99,7 +99,7 @@ void* kmalloc(size_t size)
|
||||
|
||||
// Found a good block, we return it
|
||||
curr->free = false;
|
||||
return (void*)((uintptr_t)curr + sizeof(struct heap_block_t));
|
||||
return (void*)((uintptr_t)curr + sizeof(struct heap_block));
|
||||
}
|
||||
// Continue browsing the list if nothing good was found yet
|
||||
curr = curr->next;
|
||||
@@ -127,11 +127,11 @@ void kfree(void* ptr)
|
||||
if (!ptr) return;
|
||||
|
||||
// Set it free!
|
||||
struct heap_block_t* block = (struct heap_block_t*)((uintptr_t)ptr - sizeof(struct heap_block_t));
|
||||
struct heap_block* block = (struct heap_block*)((uintptr_t)ptr - sizeof(struct heap_block));
|
||||
block->free = true;
|
||||
|
||||
// merge adjacent free blocks (coalescing)
|
||||
struct heap_block_t* curr = head;
|
||||
struct heap_block* curr = head;
|
||||
while (curr && curr->next) {
|
||||
if (curr->free && curr->next->free) {
|
||||
curr->size += sizeof(*curr) + curr->next->size;
|
||||
@@ -158,3 +158,32 @@ void* kalloc_stack()
|
||||
uint8_t* ptr = kmalloc(PROCESS_STACK_SIZE); // As it's out of kmalloc, stack is already mapped into kernel space
|
||||
return ptr ? ptr+PROCESS_STACK_SIZE : NULL;
|
||||
}
|
||||
|
||||
/*
|
||||
* kheap_info - Display heap info
|
||||
*
|
||||
* This function writes the size of the heap (total),
|
||||
* the number of allocated bytes, and the number of
|
||||
* free bytes to the standard output.
|
||||
*/
|
||||
void kheap_info()
|
||||
{
|
||||
uint64_t free_bytes = 0;
|
||||
struct heap_block* curr = (struct heap_block*)kheap_start;
|
||||
|
||||
while (curr) {
|
||||
if (curr->free == true) {
|
||||
free_bytes += curr->size;
|
||||
}
|
||||
curr = curr->next;
|
||||
}
|
||||
|
||||
uint64_t total = end-kheap_start;
|
||||
|
||||
printf("total=% 8u bytes (%u kB)\r\n"
|
||||
"alloc=% 8u bytes (%u kB)\r\n"
|
||||
" free=% 8u bytes (%u kB)\r\n",
|
||||
total, (total)/1000,
|
||||
total-free_bytes, (total-free_bytes)/1000,
|
||||
free_bytes, free_bytes/1000);
|
||||
}
|
||||
+19
-3
@@ -99,30 +99,46 @@ void paging_map_page(uint64_t* root_table, uint64_t virt, uint64_t phys, uint64_
|
||||
|
||||
uint64_t *pdpt, *pd, *pt;
|
||||
|
||||
// Any parent entry on a userspace mapping must also carry PTE_USER,
|
||||
// otherwise CPL3 accesses fault even if the final PTE is user.
|
||||
uint64_t parent_flags = PTE_PRESENT | PTE_WRITABLE;
|
||||
if (flags & PTE_USER) {
|
||||
parent_flags |= PTE_USER;
|
||||
}
|
||||
|
||||
// PML4
|
||||
// If the entry at index is not present, allocate enough space for it
|
||||
// then populate the entry with correct addr + flags
|
||||
if (!(root_table[pml4_i] & PTE_PRESENT)) {
|
||||
pdpt = alloc_page_table();
|
||||
root_table[pml4_i] = VIRT_TO_PHYS(pdpt) | PTE_PRESENT | PTE_WRITABLE;
|
||||
root_table[pml4_i] = VIRT_TO_PHYS(pdpt) | parent_flags;
|
||||
} else {
|
||||
pdpt = (uint64_t *)PHYS_TO_VIRT(root_table[pml4_i] & PTE_ADDR_MASK);
|
||||
if (flags & PTE_USER) {
|
||||
root_table[pml4_i] |= PTE_USER;
|
||||
}
|
||||
}
|
||||
|
||||
// PDPT: same here
|
||||
if (!(pdpt[pdpt_i] & PTE_PRESENT)) {
|
||||
pd = alloc_page_table();
|
||||
pdpt[pdpt_i] = VIRT_TO_PHYS(pd) | PTE_PRESENT | PTE_WRITABLE;
|
||||
pdpt[pdpt_i] = VIRT_TO_PHYS(pd) | parent_flags;
|
||||
} else {
|
||||
pd = (uint64_t *)PHYS_TO_VIRT(pdpt[pdpt_i] & PTE_ADDR_MASK);
|
||||
if (flags & PTE_USER) {
|
||||
pdpt[pdpt_i] |= PTE_USER;
|
||||
}
|
||||
}
|
||||
|
||||
// PD: and here
|
||||
if (!(pd[pd_i] & PTE_PRESENT)) {
|
||||
pt = alloc_page_table();
|
||||
pd[pd_i] = VIRT_TO_PHYS(pt) | PTE_PRESENT | PTE_WRITABLE;
|
||||
pd[pd_i] = VIRT_TO_PHYS(pt) | parent_flags;
|
||||
} else {
|
||||
pt = (uint64_t *)PHYS_TO_VIRT(pd[pd_i] & PTE_ADDR_MASK);
|
||||
if (flags & PTE_USER) {
|
||||
pd[pd_i] |= PTE_USER;
|
||||
}
|
||||
}
|
||||
|
||||
// PT: finally, populate the page table entry
|
||||
|
||||
@@ -13,6 +13,7 @@ in a specified virtual space
|
||||
compared to the PMM which allocs/frees 4kb frames ("physical pages").
|
||||
*/
|
||||
|
||||
#include "config.h"
|
||||
#include <mem/vmm.h>
|
||||
#include <mem/paging.h>
|
||||
#include <stddef.h>
|
||||
@@ -225,6 +226,43 @@ void* vmm_alloc_region(uint64_t* pml4, size_t pages, uint64_t flags)
|
||||
return NULL;
|
||||
}
|
||||
|
||||
/*
|
||||
* vmm_map_user_stack - Map a user stack
|
||||
* @pml4: the user process's PML4
|
||||
*
|
||||
* This function maps and allocates a userspace
|
||||
* stack in the user @pml4 provided, according
|
||||
* to constants USER_STACK_TOP and USER_STACK_PAGES.
|
||||
*
|
||||
* Return:
|
||||
* <addr> - User stack top address
|
||||
*/
|
||||
|
||||
uintptr_t vmm_alloc_user_stack(uint64_t* pml4)
|
||||
{
|
||||
uintptr_t stack_top = USER_STACK_TOP;
|
||||
size_t stack_size = USER_STACK_PAGES*PAGE_SIZE;
|
||||
|
||||
for (size_t i=stack_top; i>stack_top-stack_size; i-=PAGE_SIZE) {
|
||||
vmm_map(pml4, i, PTE_PRESENT | PTE_WRITABLE | PTE_USER);
|
||||
}
|
||||
return stack_top;
|
||||
}
|
||||
|
||||
uintptr_t vmm_alloc_user_code(uint64_t* pml4, void* code_addr, uint64_t code_size)
|
||||
{
|
||||
uintptr_t code_start = USER_CODE_START;
|
||||
|
||||
// Round code_size up to next page boundary
|
||||
uint64_t code_size_aligned = (code_size + PAGE_SIZE - 1) & ~(PAGE_SIZE - 1);
|
||||
|
||||
for (uint64_t i=code_start; i<code_start+code_size_aligned; i+=PAGE_SIZE) {
|
||||
vmm_map(pml4, i, PTE_PRESENT | PTE_WRITABLE | PTE_USER);
|
||||
}
|
||||
|
||||
return code_start;
|
||||
}
|
||||
|
||||
void vmm_init()
|
||||
{
|
||||
// NO U
|
||||
|
||||
+97
-14
@@ -4,20 +4,23 @@
|
||||
* @license GPL-3.0-only
|
||||
*/
|
||||
|
||||
#include "mem/paging.h"
|
||||
#include "mem/vmm.h"
|
||||
#include <stddef.h>
|
||||
#include <sched/process.h>
|
||||
#include <mem/kheap.h>
|
||||
#include <kernel.h>
|
||||
#include <string/string.h>
|
||||
#include <mem/gdt.h>
|
||||
#include <arch/gdt.h>
|
||||
#include <config.h>
|
||||
#include <io/serial/serial.h>
|
||||
|
||||
#include <io/term/flanterm.h>
|
||||
#include <mem/utils.h>
|
||||
|
||||
extern struct flanterm_context* ft_ctx;
|
||||
|
||||
struct process_t* processes_list;
|
||||
struct process_t* current_process;
|
||||
struct process* processes_list;
|
||||
struct process* current_process;
|
||||
|
||||
extern uint64_t *kernel_pml4;
|
||||
|
||||
@@ -39,10 +42,10 @@ void process_init()
|
||||
* This function prints the linked list of processes
|
||||
* to the DEBUG output.
|
||||
*/
|
||||
void process_display_list(struct process_t* processes_list)
|
||||
void process_display_list(struct process* processes_list)
|
||||
{
|
||||
int process_view_id = 0;
|
||||
struct process_t* tmp = processes_list;
|
||||
struct process* tmp = processes_list;
|
||||
while (tmp != NULL) {
|
||||
DEBUG("{%d: %p} -> ", process_view_id, tmp);
|
||||
tmp = tmp->next;
|
||||
@@ -64,11 +67,11 @@ void process_display_list(struct process_t* processes_list)
|
||||
* Return:
|
||||
* <proc> - pointer to created process
|
||||
*/
|
||||
struct process_t* process_create(char* name, void(*function)(void*), void* arg)
|
||||
struct process* process_create(char* name, void(*function)(void*), void* arg)
|
||||
{
|
||||
CLEAR_INTERRUPTS;
|
||||
struct process_t* proc = (struct process_t*)kmalloc(sizeof(struct process_t));
|
||||
struct cpu_status_t* ctx = (struct cpu_status_t*)kmalloc(sizeof(struct cpu_status_t));
|
||||
struct process* proc = (struct process*)kmalloc(sizeof(struct process));
|
||||
struct cpu_status* ctx = (struct cpu_status*)kmalloc(sizeof(struct cpu_status));
|
||||
|
||||
// No more memory?
|
||||
if (!proc) return NULL;
|
||||
@@ -95,6 +98,8 @@ struct process_t* process_create(char* name, void(*function)(void*), void* arg)
|
||||
// Kernel PML4 as it already maps code/stack (when switching to userland we'll have to change that)
|
||||
proc->root_page_table = kernel_pml4;
|
||||
|
||||
proc->kernel_stack = kalloc_stack();
|
||||
|
||||
proc->next = 0;
|
||||
|
||||
process_add(&processes_list, proc);
|
||||
@@ -108,7 +113,7 @@ struct process_t* process_create(char* name, void(*function)(void*), void* arg)
|
||||
* @processes_list: pointer to the head of the linked list
|
||||
* @process: process to add at the end of the linked list
|
||||
*/
|
||||
void process_add(struct process_t** processes_list, struct process_t* process)
|
||||
void process_add(struct process** processes_list, struct process* process)
|
||||
{
|
||||
if (!process) return;
|
||||
process->next = NULL;
|
||||
@@ -119,7 +124,7 @@ void process_add(struct process_t** processes_list, struct process_t* process)
|
||||
return;
|
||||
}
|
||||
|
||||
struct process_t* tmp = *processes_list;
|
||||
struct process* tmp = *processes_list;
|
||||
while (tmp->next != NULL) {
|
||||
tmp = tmp->next;
|
||||
}
|
||||
@@ -132,7 +137,7 @@ void process_add(struct process_t** processes_list, struct process_t* process)
|
||||
* @processes_list: pointer to head of linked list
|
||||
* @process: the process to delete from the list
|
||||
*/
|
||||
void process_delete(struct process_t** processes_list, struct process_t* process)
|
||||
void process_delete(struct process** processes_list, struct process* process)
|
||||
{
|
||||
if (!processes_list || !*processes_list || !process) return;
|
||||
|
||||
@@ -144,7 +149,7 @@ void process_delete(struct process_t** processes_list, struct process_t* process
|
||||
return;
|
||||
}
|
||||
|
||||
struct process_t* tmp = *processes_list;
|
||||
struct process* tmp = *processes_list;
|
||||
while (tmp->next && tmp->next != process) {
|
||||
tmp = tmp->next;
|
||||
}
|
||||
@@ -167,7 +172,7 @@ void process_delete(struct process_t** processes_list, struct process_t* process
|
||||
* Return:
|
||||
* <process->next> - process right after the one specified
|
||||
*/
|
||||
struct process_t* process_get_next(struct process_t* process)
|
||||
struct process* process_get_next(struct process* process)
|
||||
{
|
||||
if (!process) return NULL;
|
||||
return process->next;
|
||||
@@ -197,3 +202,81 @@ void process_exit()
|
||||
asm("hlt");
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* process_jump_to_user - Jump to userland
|
||||
* @stack_top: Address of the top of the user stack
|
||||
* @user_code: Address of the first instruction of user code
|
||||
*/
|
||||
void process_jump_to_user(uintptr_t stack_top, uintptr_t user_code)
|
||||
{
|
||||
// 0x20 | 3 = 0x23 (user data segment | 3)
|
||||
// 0x18 | 3 = 0x1B (user code segment | 3)
|
||||
asm volatile(" \
|
||||
push $0x23 \n\
|
||||
push %0 \n\
|
||||
push $0x202 \n\
|
||||
push $0x1B \n\
|
||||
push %1 \n\
|
||||
iretq \n\
|
||||
" :: "r"(stack_top), "r"(user_code));
|
||||
}
|
||||
|
||||
// Kernel stack used for interrupts from userland process.
|
||||
// Should be set in TSS.RSP0 when switching to userland process.
|
||||
uint8_t interrupt_stack[0x8000];
|
||||
|
||||
extern struct tss tss;
|
||||
|
||||
/*
|
||||
* process_create_user - Create a new user process
|
||||
* @file: pointer to Limine file structure
|
||||
* @name: name for the new process
|
||||
*
|
||||
* This function takes a loaded Limine executable
|
||||
* module, and maps its code, a user stack, sets the
|
||||
* TSS RSP0 for interrupts, and finally jumps to the
|
||||
* user code.
|
||||
*/
|
||||
void process_create_user(struct limine_file* file, char* name)
|
||||
{
|
||||
CLEAR_INTERRUPTS;
|
||||
struct process* proc = (struct process*)kmalloc(sizeof(struct process));
|
||||
struct cpu_status* ctx = (struct cpu_status*)kmalloc(sizeof(struct cpu_status));
|
||||
|
||||
if (!proc || !ctx) panic(NULL, "out of memory while creating user process");
|
||||
|
||||
strncpy(proc->name, name, PROCESS_NAME_MAX);
|
||||
memset(ctx, 0, sizeof(struct cpu_status)); // set GP registers to zero
|
||||
proc->pid = next_free_pid++;
|
||||
proc->status = READY;
|
||||
proc->next = 0;
|
||||
proc->context = ctx;
|
||||
proc->context->iret_ss = USER_DATA_SEGMENT | 3;
|
||||
proc->context->iret_cs = USER_CODE_SEGMENT | 3;
|
||||
proc->context->iret_flags = 0x202; // Interrupt Flag set
|
||||
|
||||
void* exec_addr = file->address;
|
||||
uint64_t exec_size = file->size;
|
||||
|
||||
uint64_t* user_pml4 = vmm_create_address_space();
|
||||
if (!user_pml4) panic(NULL, "failed to create user address space");
|
||||
proc->root_page_table = user_pml4;
|
||||
|
||||
uintptr_t stack_top = vmm_alloc_user_stack(user_pml4);
|
||||
uint64_t code = vmm_alloc_user_code(user_pml4, exec_addr, exec_size);
|
||||
|
||||
proc->context->iret_rsp = stack_top;
|
||||
proc->context->iret_rip = code;
|
||||
proc->kernel_stack = kalloc_stack();
|
||||
if (!proc->kernel_stack) panic(NULL, "failed to allocate kernel stack");
|
||||
|
||||
// Copy code into user pages; for that we need to temporarily switch to the user pml4
|
||||
load_cr3(VIRT_TO_PHYS((uint64_t)user_pml4));
|
||||
memcpy((uint64_t*)code, exec_addr, exec_size);
|
||||
load_cr3(VIRT_TO_PHYS((uint64_t)kernel_pml4));
|
||||
|
||||
process_add(&processes_list, proc);
|
||||
DEBUG("user process '%s' (pid=%u) enqueued for scheduling", name, proc->pid);
|
||||
SET_INTERRUPTS;
|
||||
}
|
||||
+41
-23
@@ -9,10 +9,13 @@
|
||||
#include <mem/paging.h>
|
||||
#include <stdint.h>
|
||||
#include <io/serial/serial.h>
|
||||
#include <arch/gdt.h>
|
||||
|
||||
extern struct process_t* processes_list;
|
||||
extern struct process_t* current_process;
|
||||
extern struct process_t* idle_proc;
|
||||
extern struct process* processes_list;
|
||||
extern struct process* current_process;
|
||||
extern struct process* idle_proc;
|
||||
|
||||
extern struct tss tss;
|
||||
|
||||
/*
|
||||
* scheduler_init - Choose the first process
|
||||
@@ -20,6 +23,7 @@ extern struct process_t* idle_proc;
|
||||
void scheduler_init()
|
||||
{
|
||||
current_process = processes_list;
|
||||
DEBUG("scheduler starting with: pid=%u, name='%s', context=%p", current_process->pid, current_process->name, current_process->context);
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -32,49 +36,63 @@ void scheduler_init()
|
||||
* Return:
|
||||
* <context> - CPU context for next process
|
||||
*/
|
||||
struct cpu_status_t* scheduler_schedule(struct cpu_status_t* context)
|
||||
struct cpu_status* scheduler_schedule(struct cpu_status* context)
|
||||
{
|
||||
if (context == NULL) {
|
||||
panic(NULL, "Scheduler called with NULL context");
|
||||
}
|
||||
|
||||
if (current_process == NULL) {
|
||||
// If no more processes, then set IDLE as the current process, that's it.
|
||||
current_process = idle_proc;
|
||||
panic(NULL, "current_process is NULL");
|
||||
}
|
||||
|
||||
if (current_process == idle_proc && current_process->next == NULL)
|
||||
{
|
||||
return idle_proc->context;
|
||||
if (current_process->context == NULL) {
|
||||
panic(NULL, "current_process->context is NULL");
|
||||
}
|
||||
|
||||
current_process->context = context;
|
||||
|
||||
for (;;) {
|
||||
struct process_t* prev_process = current_process;
|
||||
if (current_process->next != NULL) {
|
||||
if (current_process->status == DEAD) {
|
||||
struct process* dead_process = current_process;
|
||||
struct process* next_process = (dead_process->next != NULL) ? dead_process->next : processes_list;
|
||||
|
||||
process_delete(&processes_list, dead_process);
|
||||
|
||||
if (processes_list == NULL || next_process == dead_process) {
|
||||
current_process = idle_proc;
|
||||
return idle_proc->context;
|
||||
}
|
||||
|
||||
current_process = next_process;
|
||||
} else if (current_process->next != NULL) {
|
||||
current_process = current_process->next;
|
||||
} else {
|
||||
current_process = processes_list;
|
||||
}
|
||||
|
||||
if (current_process != NULL && current_process->status == DEAD) {
|
||||
process_delete(&prev_process, current_process);
|
||||
current_process = NULL;
|
||||
for (;;) {
|
||||
if (current_process->status == DEAD) {
|
||||
struct process* dead_process = current_process;
|
||||
struct process* next_process = (current_process->next != NULL) ? current_process->next : processes_list;
|
||||
|
||||
process_delete(&processes_list, dead_process);
|
||||
|
||||
if (processes_list == NULL || next_process == dead_process) {
|
||||
current_process = idle_proc;
|
||||
return idle_proc->context;
|
||||
} else {
|
||||
}
|
||||
|
||||
current_process = next_process;
|
||||
continue;
|
||||
}
|
||||
|
||||
current_process->status = RUNNING;
|
||||
/* if (prev_process != current_process) {
|
||||
DEBUG("Changed from {pid=%u, name=%s} to {pid=%u, name=%s}", prev_process->pid, prev_process->name, current_process->pid, current_process->name);
|
||||
} */
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
//DEBUG("current_process={pid=%u, name='%s', root_page_table[virt]=%p}", current_process->pid, current_process->name, current_process->root_page_table);
|
||||
|
||||
// Here, we chose next running process so we load its kernel stack & page tables
|
||||
tss.rsp0 = (uint64_t)current_process->kernel_stack;
|
||||
load_cr3(VIRT_TO_PHYS((uint64_t)current_process->root_page_table));
|
||||
//DEBUG("Loaded process PML4 into CR3");
|
||||
|
||||
return current_process->context;
|
||||
}
|
||||
@@ -16,7 +16,7 @@
|
||||
* Saves the RFLAGS register, then acquires a lock.
|
||||
* Pause instruction is used to ease the CPU.
|
||||
*/
|
||||
void spinlock_acquire(struct spinlock_t* lock)
|
||||
void spinlock_acquire(struct spinlock* lock)
|
||||
{
|
||||
uint64_t rflags;
|
||||
asm volatile("pushfq ; pop %0 ; cli" : "=rm"(rflags) : : "memory");
|
||||
@@ -36,7 +36,7 @@ void spinlock_acquire(struct spinlock_t* lock)
|
||||
* unlocks it (clears locked state).
|
||||
* RFLAGS is then restored.
|
||||
*/
|
||||
void spinlock_release(struct spinlock_t* lock)
|
||||
void spinlock_release(struct spinlock* lock)
|
||||
{
|
||||
uint64_t rflags = lock->rflags;
|
||||
__atomic_clear(&lock->locked, __ATOMIC_RELEASE);
|
||||
|
||||
@@ -0,0 +1,21 @@
|
||||
bits 64
|
||||
|
||||
section .data
|
||||
hi db "hi from userland :) we did it man", 0x0A, 0x0d, 0
|
||||
|
||||
section .text
|
||||
|
||||
hello:
|
||||
mov rax, 0x1 ;sys_write
|
||||
mov rdi, 0x1 ;stdout
|
||||
lea rsi, [rel hi] ;char* buf
|
||||
mov rdx, 35 ;count
|
||||
int 0x80
|
||||
|
||||
.end:
|
||||
mov rax, 0x3C ;sys_exit
|
||||
mov rdi, 0x0 ;error_code
|
||||
int 0x80
|
||||
|
||||
.loop:
|
||||
jmp .loop
|
||||
@@ -0,0 +1,25 @@
|
||||
bits 64
|
||||
|
||||
section .data
|
||||
hello db 0x0A, 0x0D, "User program 2 speaking", 0x0A, 0x0D, 0
|
||||
|
||||
section .text
|
||||
|
||||
_start:
|
||||
mov rax, 0x1 ;sys_write
|
||||
mov rdi, 0x1 ;stdout
|
||||
lea rsi, [rel hello]
|
||||
mov rdx, 27 ;count
|
||||
int 0x80
|
||||
|
||||
; when we are ready to have an os specific toolchain,
|
||||
; this bit (exit & loop) should be appended at the end of every
|
||||
; C program we compile.
|
||||
|
||||
.end:
|
||||
mov rax, 0x3C
|
||||
mov rdi, 0x0
|
||||
int 0x80
|
||||
|
||||
.loop:
|
||||
jmp .loop
|
||||
Reference in New Issue
Block a user