I tried to explain everything in detail and to keep the code as simple as possible. If you have any questions, suggestions, or issues, please leave a comment or on Github. The source code is available in a repository, too.

To avoid bugs and strange errors on old CPUs we should check if the processor supports every needed feature. If not, the kernel should abort and display an error message. To handle errors easily, we create an error procedure in boot.asm. It prints a rudimentary ERR: X message, where X is an error code letter, and hangs:

At address 0xb8000 begins the so-called VGA text buffer. It’s an array of screen characters that are displayed by the graphics card. A will cover the VGA buffer in detail and create a Rust interface to it. But for now, manual bit-fiddling is the easiest option.

A screen character consists of a 8 bit color code and a 8 bit ASCII character. We used the color code 4f for all characters, which means white text on red background. 0x52 is an ASCII R, 0x45 is an E, 0x3a is a :, and 0x20 is a space. The second space is overwritten by the given ASCII byte. Finally the CPU is stopped with the hlt instruction.

Now we can add some check functions. A function is just a normal label with an ret (return) instruction at the end. The call instruction can be used to call it. Unlike the jmp instruction that just jumps to a memory address, the call instruction will push a return address to the stack (and the ret will jump to this address). But we don’t have a stack yet. The in the esp register could point to some important data or even invalid memory. So we need to update it and point it to some valid stack memory.

To create stack memory we reserve some bytes at the end of our boot.asm:

  1. ...
  2. section .bss
  3. stack_bottom:
  4. resb 64
  5. stack_top:

A stack doesn’t need to be initialized because we will pop only when we pushed before. So storing the stack memory in the executable file would make it unnecessary large. By using the section and the resb (reserve byte) command, we just store the length of the uninitialized data (= 64). When loading the executable, GRUB will create the section of required size in memory.

To use the new stack, we update the stack pointer register right after start:

  1. global start
  2. section .text
  3. bits 32
  4. start:
  5. mov esp, stack_top
  6. ; print `OK` to screen
  7. ...

We use stack_top because the stack grows downwards: A push eax subtracts 4 from esp and does a mov [esp], eax afterwards (eax is a general purpose register).

Now we have a valid stack pointer and are able to call functions. The following check functions are just here for completeness and I won’t explain details. Basically they all work the same: They will check for a feature and jump to error if it’s not available.

Multiboot check

We rely on some Multiboot features in the next posts. To make sure the kernel was really loaded by a Multiboot compliant bootloader, we can check the eax register. According to the Multiboot specification (), the bootloader must write the magic value 0x36d76289 to it before loading a kernel. To verify that we can add a simple function:

  1. check_multiboot:
  2. cmp eax, 0x36d76289
  3. jne .no_multiboot
  4. ret
  5. .no_multiboot:
  6. mov al, "0"
  7. jmp error

We use the cmp instruction to compare the value in eax to the magic value. If the values are equal, the cmp instruction sets the zero flag in the FLAGS register. The jne (“jump if not equal”) instruction reads this zero flag and jumps to the given address if it’s not set. Thus we jump to the .no_multiboot label if eax does not contain the magic value.

In no_multiboot, we use the jmp (“jump”) instruction to jump to our error function. We could just as well use the call instruction, which additionally pushes the return address. But the return address is not needed because error never returns. To pass 0 as error code to the error function, we move it into al before the jump (error will read it from there).

CPUID is a CPU instruction that can be used to get various information about the CPU. But not every processor supports it. CPUID detection is quite laborious, so we just copy a detection function from the :

  1. check_cpuid:
  2. ; Check if CPUID is supported by attempting to flip the ID bit (bit 21)
  3. ; in the FLAGS register. If we can flip it, CPUID is available.
  4. ; Copy FLAGS in to EAX via stack
  5. pushfd
  6. pop eax
  7. ; Copy to ECX as well for comparing later on
  8. mov ecx, eax
  9. ; Flip the ID bit
  10. xor eax, 1 << 21
  11. ; Copy EAX to FLAGS via the stack
  12. push eax
  13. popfd
  14. ; Copy FLAGS back to EAX (with the flipped bit if CPUID is supported)
  15. pushfd
  16. pop eax
  17. ; Restore FLAGS from the old version stored in ECX (i.e. flipping the
  18. ; ID bit back if it was ever flipped).
  19. push ecx
  20. popfd
  21. ; Compare EAX and ECX. If they are equal then that means the bit
  22. ; wasn't flipped, and CPUID isn't supported.
  23. cmp eax, ecx
  24. je .no_cpuid
  25. ret
  26. .no_cpuid:
  27. mov al, "1"
  28. jmp error

Basically, the CPUID instruction is supported if we can flip some bit in the FLAGS register. We can’t operate on the flags register directly, so we need to load it into some general purpose register such as eax first. The only way to do this is to push the FLAGS register on the stack through the pushfd instruction and then pop it into eax. Equally, we write it back through push ecx and popfd. To flip the bit we use the xor instruction to perform an . Finally we compare the two values and jump to .no_cpuid if both are equal (je – “jump if equal”). The .no_cpuid code just jumps to the error function with error code 1.

Don’t worry, you don’t need to understand the details.

Long Mode check

Now we can use CPUID to detect whether long mode can be used. I use code from again:

  1. ; test if extended processor info in available
  2. mov eax, 0x80000000 ; implicit argument for cpuid
  3. cpuid ; get highest supported argument
  4. jb .no_long_mode ; if it's less, the CPU is too old for long mode
  5. ; use extended info to test if long mode is available
  6. mov eax, 0x80000001 ; argument for extended processor info
  7. cpuid ; returns various feature bits in ecx and edx
  8. test edx, 1 << 29 ; test if the LM-bit is set in the D-register
  9. jz .no_long_mode ; If it's not set, there is no long mode
  10. ret
  11. .no_long_mode:
  12. mov al, "2"
  13. jmp error

Like many low-level things, CPUID is a bit strange. Instead of taking a parameter, the cpuid instruction implicitly uses the eax register as argument. To test if long mode is available, we need to call cpuid with 0x80000001 in eax. This loads some information to the ecx and edx registers. Long mode is supported if the 29th bit in edx is set. Wikipedia has detailed information.

If you look at the assembly above, you’ll probably notice that we call cpuid twice. The reason is that the CPUID command started with only a few functions and was extended over time. So old processors may not know the 0x80000001 argument at all. To test if they do, we need to invoke cpuid with 0x80000000 in eax first. It returns the highest supported parameter value in eax. If it’s at least 0x80000001, we can test for long mode as described above. Else the CPU is old and doesn’t know what long mode is either. In that case, we directly jump to .no_long_mode through the jb instruction (“jump if below”).

We just call these check functions right after start:

When the CPU doesn’t support a needed feature, we get an error message with an unique error code. Now we can start the real work.

In long mode, x86 uses a page size of 4096 bytes and a 4 level page table that consists of:

  • the Page-Map Level-4 Table (PML4),
  • the Page-Directory Pointer Table (PDP),
  • the Page-Directory Table (PD),
  • and the Page Table (PT).

As I don’t like these names, I will call them P4, P3, P2, and P1 from now on.

Each page table contains 512 entries and one entry is 8 bytes, so they fit exactly in one page (512*8 = 4096). To translate a virtual address to a physical address the CPU[^hardware_lookup] will do the following[^virtual_physical_translation_source]:

  1. Get the address of the P4 table from the CR3 register
  2. Use bits 39-47 (9 bits) as an index into P4 (2^9 = 512 = number of entries)
  3. Use the following 9 bits as an index into P3
  4. Use the following 9 bits as an index into P2
  5. Use the following 9 bits as an index into P1
  6. Use the last 12 bits as page offset (2^12 = 4096 = page size)

But what happens to bits 48-63 of the 64-bit virtual address? Well, they can’t be used. The “64-bit” long mode is in fact just a 48-bit mode. The bits 48-63 must be copies of bit 47, so each valid virtual address is still unique. For more information see .

An entry in the P4, P3, P2, and P1 tables consists of the page aligned 52-bit physical address of the frame or the next page table and the following bits that can be OR-ed in:

Set Up Identity Paging

When we switch to long mode, paging will be activated automatically. The CPU will then try to read the instruction at the following address, but this address is now a virtual address. So we need to do identity mapping, i.e. map a physical address to the same virtual address.

The huge page bit is now very useful to us. It creates a 2MiB (when used in P2) or even a 1GiB page (when used in P3). So we could map the first gigabytes of the kernel with only one P4 and one P3 table by using 1GiB pages. Unfortunately 1GiB pages are relatively new feature, for example Intel introduced it 2010 in the . Therefore we will use 2MiB pages instead to make our kernel compatible to older computers, too.

To identity map the first gigabyte of our kernel with 512 2MiB pages, we need one P4, one P3, and one P2 table. Of course we will replace them with finer-grained tables later. But now that we’re stuck with assembly, we choose the easiest way.

We can add these two tables at the beginning[^page_table_alignment] of the .bss section:

  1. ...
  2. section .bss
  3. align 4096
  4. p4_table:
  5. resb 4096
  6. p3_table:
  7. resb 4096
  8. p2_table:
  9. resb 4096
  10. stack_bottom:
  11. resb 64
  12. stack_top:

The resb command reserves the specified amount of bytes without initializing them, so the 8KiB don’t need to be saved in the executable. The align 4096 ensures that the page tables are page aligned.

When GRUB creates the .bss section in memory, it will initialize it to 0. So the p4_table is already valid (it contains 512 non-present entries) but not very useful. To be able to map 2MiB pages, we need to link P4’s first entry to the p3_table and P3’s first entry to the the p2_table:

  1. set_up_page_tables:
  2. ; map first P4 entry to P3 table
  3. mov eax, p3_table
  4. or eax, 0b11 ; present + writable
  5. mov [p4_table], eax
  6. ; map first P3 entry to P2 table
  7. mov eax, p2_table
  8. or eax, 0b11 ; present + writable
  9. mov [p3_table], eax
  10. ; TODO map each P2 entry to a huge 2MiB page
  11. ret

We just set the present and writable bits (0b11 is a binary number) in the aligned P3 table address and move it to the first 4 bytes of the P4 table. Then we do the same to link the first P3 entry to the p2_table.

Now we need to map P2’s first entry to a huge page starting at 0, P2’s second entry to a huge page starting at 2MiB, P2’s third entry to a huge page starting at 4MiB, and so on. It’s time for our first (and only) assembly loop:

  1. set_up_page_tables:
  2. ...
  3. ; map each P2 entry to a huge 2MiB page
  4. mov ecx, 0 ; counter variable
  5. .map_p2_table:
  6. ; map ecx-th P2 entry to a huge page that starts at address 2MiB*ecx
  7. mov eax, 0x200000 ; 2MiB
  8. mul ecx ; start address of ecx-th page
  9. or eax, 0b10000011 ; present + writable + huge
  10. mov [p2_table + ecx * 8], eax ; map ecx-th entry
  11. inc ecx ; increase counter
  12. cmp ecx, 512 ; if counter == 512, the whole P2 table is mapped
  13. jne .map_p2_table ; else map the next entry
  14. ret

Maybe I should first explain how an assembly loop works. We use the ecx register as a counter variable, just like i in a for loop. After mapping the ecx-th entry, we increase ecx by one and jump to .map_p2_table again if it’s still smaller than 512.

To map a P2 entry we first calculate the start address of its page in eax: The ecx-th entry needs to be mapped to ecx * 2MiB. We use the mul operation for that, which multiplies eax with the given register and stores the result in eax. Then we set the present, writable, and huge page bits and write it to the P2 entry. The address of the ecx-th entry in P2 is p2_table + ecx * 8, because each entry is 8 bytes large.

Now the first gigabyte (512 * 2MiB) of our kernel is identity mapped and thus accessible through the same physical and virtual addresses.

To enable paging and enter long mode, we need to do the following:

  1. write the address of the P4 table to the CR3 register (the CPU will look there, see the )
  2. long mode is an extension of Physical Address Extension (PAE), so we need to enable PAE first
  3. Set the long mode bit in the EFER register
  4. Enable Paging

The assembly function looks like this (some boring bit-moving to various registers):

  1. enable_paging:
  2. ; load P4 to cr3 register (cpu uses this to access the P4 table)
  3. mov eax, p4_table
  4. mov cr3, eax
  5. mov eax, cr4
  6. or eax, 1 << 5
  7. ; set the long mode bit in the EFER MSR (model specific register)
  8. mov ecx, 0xC0000080
  9. rdmsr
  10. or eax, 1 << 8
  11. wrmsr
  12. ; enable paging in the cr0 register
  13. mov eax, cr0
  14. or eax, 1 << 31
  15. mov cr0, eax
  16. ret

The or eax, 1 << X is a common pattern. It sets the bit X in the eax register (<< is a left shift). Through rdmsr and wrmsr it’s possible to read/write to the so-called model specific registers at address ecx (in this case ecx points to the EFER register).

Finally we need to call our new functions in start:

  1. ...
  2. start:
  3. mov esp, stack_top
  4. call check_multiboot
  5. call check_cpuid
  6. call check_long_mode
  7. call set_up_page_tables ; new
  8. call enable_paging ; new
  9. ; print `OK` to screen
  10. mov dword [0xb8000], 0x2f4b2f4f
  11. hlt
  12. ...

After enabling Paging, the processor is in long mode. So we can use 64-bit instructions now, right? Wrong. The processor is still in a 32-bit compatibility submode. To actually execute 64-bit code, we need to set up a new Global Descriptor Table. The Global Descriptor Table (GDT) was used for Segmentation in old operating systems. I won’t explain Segmentation but the Three Easy Pieces OS book has good introduction () again.

Today almost everyone uses Paging instead of Segmentation (and so do we). But on x86, a GDT is always required, even when you’re not using Segmentation. GRUB has set up a valid 32-bit GDT for us but now we need to switch to a long mode GDT.

A GDT always starts with a 0-entry and contains an arbitrary number of segment entries afterwards. A 64-bit entry has the following format:

Bit(s) Name Meaning
0-41 ignored ignored in 64-bit mode
42 conforming the current privilege level can be higher than the specified level for code segments (else it must match exactly)
43 executable if set, it’s a code segment, else it’s a data segment
44 descriptor type should be 1 for code and data segments
45-46 privilege the ring level: 0 for kernel, 3 for user
47 present must be 1 for valid selectors
48-52 ignored ignored in 64-bit mode
53 64-bit should be set for 64-bit code segments
54 32-bit must be 0 for 64-bit segments
55-63 ignored ignored in 64-bit mode

We need one code segment, a data segment is not necessary in 64-bit mode. Code segments have the following bits set: descriptor type, present, executable and the 64-bit flag. Translated to assembly the long mode GDT looks like this:

We chose the .rodata section here because it’s initialized read-only data. The dq command stands for define quad and outputs a 64-bit constant (similar to dw and dd). And the (1<<43) is a bit shift that sets bit 43.

Loading the GDT

To load our new 64-bit GDT, we have to tell the CPU its address and length. We do this by passing the memory location of a special pointer structure to the lgdt (load GDT) instruction. The pointer structure looks like this:

  1. gdt64:
  2. dq 0 ; zero entry
  3. dq (1<<43) | (1<<44) | (1<<47) | (1<<53) ; code segment
  4. .pointer:
  5. dw $ - gdt64 - 1
  6. dq gdt64

The first 2 bytes specify the (GDT length - 1). The $ is a special symbol that is replaced with the current address (it’s equal to .pointer in our case). The following 8 bytes specify the GDT address. Labels that start with a point (such as .pointer) are sub-labels of the last label without point. To access them, they must be prefixed with the parent label (e.g., gdt64.pointer).

Now we can load the GDT in start:

  1. start:
  2. ...
  3. call enable_paging
  4. ; load the 64-bit GDT
  5. lgdt [gdt64.pointer]
  6. ; print `OK` to screen
  7. ...

When you still see the green OK, everything went fine and the new GDT is loaded. But we still can’t execute 64-bit code: The code selector register cs still has the values from the old GDT. To update it, we need to load it with the GDT offset (in bytes) of the desired segment. In our case the code segment starts at byte 8 of the GDT, but we don’t want to hardcode that 8 (in case we modify our GDT later). Instead, we add a .code label to our GDT, that calculates the offset directly from the GDT:

  1. section .rodata
  2. gdt64:
  3. dq 0 ; zero entry
  4. .code: equ $ - gdt64 ; new
  5. dq (1<<43) | (1<<44) | (1<<47) | (1<<53) ; code segment
  6. .pointer:
  7. ...

We can’t just use a normal label here, since we need the table offset. We calculate this offset using the current address $ and set the label to this value using equ. Now we can use gdt64.code instead of 8 and this label will still work if we modify the GDT.

In order to finally enter the true 64-bit mode, we need to load cs with gdt64.code. But we can’t do it through mov. The only way to reload the code selector is a far jump or a far return. These instructions work like a normal jump/return but change the code selector. We use a far jump to a long mode label:

  1. global start
  2. extern long_mode_start
  3. ...
  4. start:
  5. ...
  6. lgdt [gdt64.pointer]
  7. jmp gdt64.code:long_mode_start
  8. ...

The actual long_mode_start label is defined as extern, so it’s part of another file. The jmp gdt64.code:long_mode_start is the mentioned far jump.

I put the 64-bit code into a new file to separate it from the 32-bit code, thereby we can’t call the (now invalid) 32-bit code accidentally. The new file (I named it long_mode_init.asm) looks like this:

  1. global long_mode_start
  2. section .text
  3. bits 64
  4. long_mode_start:
  5. ; print `OKAY` to screen
  6. mov rax, 0x2f592f412f4b2f4f
  7. mov qword [0xb8000], rax
  8. hlt

You should see a green OKAY on the screen. Some notes on this last step:

  • As the CPU expects 64-bit instructions now, we use bits 64
  • We can now use the extended registers. Instead of the 32-bit eax, ebx, etc. we now have the 64-bit rax, rbx, …

Congratulations! You have successfully wrestled through this CPU configuration and compatibility mode mess :).

One Last Thing

Above, we reloaded the code segment register cs with the new GDT offset. However, the data segment registers ss, ds, es, fs, and still contain the data segment offsets of the old GDT. This isn’t necessarily bad, since they’re ignored by almost all instructions in 64-bit mode. However, there are a few instructions that expect a valid data segment descriptor or the null descriptor in those registers. An example is the the iretq instruction that we’ll need in the post.

To avoid future problems, we reload all data segment registers with null:

It’s time to finally leave assembly behind and switch to . Rust is a systems language without garbage collections that guarantees memory safety. Through a real type system and many abstractions it feels like a high-level language but can still be low-level enough for OS development. The next post describes the Rust setup.

[^hardwarelookup]: In the x86 architecture, the page tables are _hardware walked, so the CPU will look at the table on its own when it needs a translation. Other architectures, for example MIPS, just throw an exception and let the OS translate the virtual address.

[^virtual_physical_translation_source]: Image source: Wikipedia, with modified font size, page table naming, and removed sign extended bits. The modified file is licensed under the Creative Commons Attribution-Share Alike 3.0 Unported license.

[^page_table_alignment]: Page tables need to be page-aligned as the bits 0-11 are used for flags. By putting these tables at the beginning of .bss, the linker can just page align the whole section and we don’t have unused padding bytes in between.