mirror of https://github.com/Qortal/Brooklyn
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
734 lines
16 KiB
734 lines
16 KiB
// SPDX-License-Identifier: GPL-2.0 |
|
#include <linux/init.h> |
|
#include <linux/async.h> |
|
#include <linux/fs.h> |
|
#include <linux/slab.h> |
|
#include <linux/types.h> |
|
#include <linux/fcntl.h> |
|
#include <linux/delay.h> |
|
#include <linux/string.h> |
|
#include <linux/dirent.h> |
|
#include <linux/syscalls.h> |
|
#include <linux/utime.h> |
|
#include <linux/file.h> |
|
#include <linux/memblock.h> |
|
#include <linux/mm.h> |
|
#include <linux/namei.h> |
|
#include <linux/init_syscalls.h> |
|
|
|
static ssize_t __init xwrite(struct file *file, const char *p, size_t count, |
|
loff_t *pos) |
|
{ |
|
ssize_t out = 0; |
|
|
|
/* sys_write only can write MAX_RW_COUNT aka 2G-4K bytes at most */ |
|
while (count) { |
|
ssize_t rv = kernel_write(file, p, count, pos); |
|
|
|
if (rv < 0) { |
|
if (rv == -EINTR || rv == -EAGAIN) |
|
continue; |
|
return out ? out : rv; |
|
} else if (rv == 0) |
|
break; |
|
|
|
p += rv; |
|
out += rv; |
|
count -= rv; |
|
} |
|
|
|
return out; |
|
} |
|
|
|
static __initdata char *message; |
|
static void __init error(char *x) |
|
{ |
|
if (!message) |
|
message = x; |
|
} |
|
|
|
static void panic_show_mem(const char *fmt, ...) |
|
{ |
|
va_list args; |
|
|
|
show_mem(0, NULL); |
|
va_start(args, fmt); |
|
panic(fmt, args); |
|
va_end(args); |
|
} |
|
|
|
/* link hash */ |
|
|
|
#define N_ALIGN(len) ((((len) + 1) & ~3) + 2) |
|
|
|
static __initdata struct hash { |
|
int ino, minor, major; |
|
umode_t mode; |
|
struct hash *next; |
|
char name[N_ALIGN(PATH_MAX)]; |
|
} *head[32]; |
|
|
|
static inline int hash(int major, int minor, int ino) |
|
{ |
|
unsigned long tmp = ino + minor + (major << 3); |
|
tmp += tmp >> 5; |
|
return tmp & 31; |
|
} |
|
|
|
static char __init *find_link(int major, int minor, int ino, |
|
umode_t mode, char *name) |
|
{ |
|
struct hash **p, *q; |
|
for (p = head + hash(major, minor, ino); *p; p = &(*p)->next) { |
|
if ((*p)->ino != ino) |
|
continue; |
|
if ((*p)->minor != minor) |
|
continue; |
|
if ((*p)->major != major) |
|
continue; |
|
if (((*p)->mode ^ mode) & S_IFMT) |
|
continue; |
|
return (*p)->name; |
|
} |
|
q = kmalloc(sizeof(struct hash), GFP_KERNEL); |
|
if (!q) |
|
panic_show_mem("can't allocate link hash entry"); |
|
q->major = major; |
|
q->minor = minor; |
|
q->ino = ino; |
|
q->mode = mode; |
|
strcpy(q->name, name); |
|
q->next = NULL; |
|
*p = q; |
|
return NULL; |
|
} |
|
|
|
static void __init free_hash(void) |
|
{ |
|
struct hash **p, *q; |
|
for (p = head; p < head + 32; p++) { |
|
while (*p) { |
|
q = *p; |
|
*p = q->next; |
|
kfree(q); |
|
} |
|
} |
|
} |
|
|
|
static long __init do_utime(char *filename, time64_t mtime) |
|
{ |
|
struct timespec64 t[2]; |
|
|
|
t[0].tv_sec = mtime; |
|
t[0].tv_nsec = 0; |
|
t[1].tv_sec = mtime; |
|
t[1].tv_nsec = 0; |
|
return init_utimes(filename, t); |
|
} |
|
|
|
static __initdata LIST_HEAD(dir_list); |
|
struct dir_entry { |
|
struct list_head list; |
|
char *name; |
|
time64_t mtime; |
|
}; |
|
|
|
static void __init dir_add(const char *name, time64_t mtime) |
|
{ |
|
struct dir_entry *de = kmalloc(sizeof(struct dir_entry), GFP_KERNEL); |
|
if (!de) |
|
panic_show_mem("can't allocate dir_entry buffer"); |
|
INIT_LIST_HEAD(&de->list); |
|
de->name = kstrdup(name, GFP_KERNEL); |
|
de->mtime = mtime; |
|
list_add(&de->list, &dir_list); |
|
} |
|
|
|
static void __init dir_utime(void) |
|
{ |
|
struct dir_entry *de, *tmp; |
|
list_for_each_entry_safe(de, tmp, &dir_list, list) { |
|
list_del(&de->list); |
|
do_utime(de->name, de->mtime); |
|
kfree(de->name); |
|
kfree(de); |
|
} |
|
} |
|
|
|
static __initdata time64_t mtime; |
|
|
|
/* cpio header parsing */ |
|
|
|
static __initdata unsigned long ino, major, minor, nlink; |
|
static __initdata umode_t mode; |
|
static __initdata unsigned long body_len, name_len; |
|
static __initdata uid_t uid; |
|
static __initdata gid_t gid; |
|
static __initdata unsigned rdev; |
|
|
|
static void __init parse_header(char *s) |
|
{ |
|
unsigned long parsed[12]; |
|
char buf[9]; |
|
int i; |
|
|
|
buf[8] = '\0'; |
|
for (i = 0, s += 6; i < 12; i++, s += 8) { |
|
memcpy(buf, s, 8); |
|
parsed[i] = simple_strtoul(buf, NULL, 16); |
|
} |
|
ino = parsed[0]; |
|
mode = parsed[1]; |
|
uid = parsed[2]; |
|
gid = parsed[3]; |
|
nlink = parsed[4]; |
|
mtime = parsed[5]; /* breaks in y2106 */ |
|
body_len = parsed[6]; |
|
major = parsed[7]; |
|
minor = parsed[8]; |
|
rdev = new_encode_dev(MKDEV(parsed[9], parsed[10])); |
|
name_len = parsed[11]; |
|
} |
|
|
|
/* FSM */ |
|
|
|
static __initdata enum state { |
|
Start, |
|
Collect, |
|
GotHeader, |
|
SkipIt, |
|
GotName, |
|
CopyFile, |
|
GotSymlink, |
|
Reset |
|
} state, next_state; |
|
|
|
static __initdata char *victim; |
|
static unsigned long byte_count __initdata; |
|
static __initdata loff_t this_header, next_header; |
|
|
|
static inline void __init eat(unsigned n) |
|
{ |
|
victim += n; |
|
this_header += n; |
|
byte_count -= n; |
|
} |
|
|
|
static __initdata char *collected; |
|
static long remains __initdata; |
|
static __initdata char *collect; |
|
|
|
static void __init read_into(char *buf, unsigned size, enum state next) |
|
{ |
|
if (byte_count >= size) { |
|
collected = victim; |
|
eat(size); |
|
state = next; |
|
} else { |
|
collect = collected = buf; |
|
remains = size; |
|
next_state = next; |
|
state = Collect; |
|
} |
|
} |
|
|
|
static __initdata char *header_buf, *symlink_buf, *name_buf; |
|
|
|
static int __init do_start(void) |
|
{ |
|
read_into(header_buf, 110, GotHeader); |
|
return 0; |
|
} |
|
|
|
static int __init do_collect(void) |
|
{ |
|
unsigned long n = remains; |
|
if (byte_count < n) |
|
n = byte_count; |
|
memcpy(collect, victim, n); |
|
eat(n); |
|
collect += n; |
|
if ((remains -= n) != 0) |
|
return 1; |
|
state = next_state; |
|
return 0; |
|
} |
|
|
|
static int __init do_header(void) |
|
{ |
|
if (memcmp(collected, "070707", 6)==0) { |
|
error("incorrect cpio method used: use -H newc option"); |
|
return 1; |
|
} |
|
if (memcmp(collected, "070701", 6)) { |
|
error("no cpio magic"); |
|
return 1; |
|
} |
|
parse_header(collected); |
|
next_header = this_header + N_ALIGN(name_len) + body_len; |
|
next_header = (next_header + 3) & ~3; |
|
state = SkipIt; |
|
if (name_len <= 0 || name_len > PATH_MAX) |
|
return 0; |
|
if (S_ISLNK(mode)) { |
|
if (body_len > PATH_MAX) |
|
return 0; |
|
collect = collected = symlink_buf; |
|
remains = N_ALIGN(name_len) + body_len; |
|
next_state = GotSymlink; |
|
state = Collect; |
|
return 0; |
|
} |
|
if (S_ISREG(mode) || !body_len) |
|
read_into(name_buf, N_ALIGN(name_len), GotName); |
|
return 0; |
|
} |
|
|
|
static int __init do_skip(void) |
|
{ |
|
if (this_header + byte_count < next_header) { |
|
eat(byte_count); |
|
return 1; |
|
} else { |
|
eat(next_header - this_header); |
|
state = next_state; |
|
return 0; |
|
} |
|
} |
|
|
|
static int __init do_reset(void) |
|
{ |
|
while (byte_count && *victim == '\0') |
|
eat(1); |
|
if (byte_count && (this_header & 3)) |
|
error("broken padding"); |
|
return 1; |
|
} |
|
|
|
static void __init clean_path(char *path, umode_t fmode) |
|
{ |
|
struct kstat st; |
|
|
|
if (!init_stat(path, &st, AT_SYMLINK_NOFOLLOW) && |
|
(st.mode ^ fmode) & S_IFMT) { |
|
if (S_ISDIR(st.mode)) |
|
init_rmdir(path); |
|
else |
|
init_unlink(path); |
|
} |
|
} |
|
|
|
static int __init maybe_link(void) |
|
{ |
|
if (nlink >= 2) { |
|
char *old = find_link(major, minor, ino, mode, collected); |
|
if (old) { |
|
clean_path(collected, 0); |
|
return (init_link(old, collected) < 0) ? -1 : 1; |
|
} |
|
} |
|
return 0; |
|
} |
|
|
|
static __initdata struct file *wfile; |
|
static __initdata loff_t wfile_pos; |
|
|
|
static int __init do_name(void) |
|
{ |
|
state = SkipIt; |
|
next_state = Reset; |
|
if (strcmp(collected, "TRAILER!!!") == 0) { |
|
free_hash(); |
|
return 0; |
|
} |
|
clean_path(collected, mode); |
|
if (S_ISREG(mode)) { |
|
int ml = maybe_link(); |
|
if (ml >= 0) { |
|
int openflags = O_WRONLY|O_CREAT; |
|
if (ml != 1) |
|
openflags |= O_TRUNC; |
|
wfile = filp_open(collected, openflags, mode); |
|
if (IS_ERR(wfile)) |
|
return 0; |
|
wfile_pos = 0; |
|
|
|
vfs_fchown(wfile, uid, gid); |
|
vfs_fchmod(wfile, mode); |
|
if (body_len) |
|
vfs_truncate(&wfile->f_path, body_len); |
|
state = CopyFile; |
|
} |
|
} else if (S_ISDIR(mode)) { |
|
init_mkdir(collected, mode); |
|
init_chown(collected, uid, gid, 0); |
|
init_chmod(collected, mode); |
|
dir_add(collected, mtime); |
|
} else if (S_ISBLK(mode) || S_ISCHR(mode) || |
|
S_ISFIFO(mode) || S_ISSOCK(mode)) { |
|
if (maybe_link() == 0) { |
|
init_mknod(collected, mode, rdev); |
|
init_chown(collected, uid, gid, 0); |
|
init_chmod(collected, mode); |
|
do_utime(collected, mtime); |
|
} |
|
} |
|
return 0; |
|
} |
|
|
|
static int __init do_copy(void) |
|
{ |
|
if (byte_count >= body_len) { |
|
struct timespec64 t[2] = { }; |
|
if (xwrite(wfile, victim, body_len, &wfile_pos) != body_len) |
|
error("write error"); |
|
|
|
t[0].tv_sec = mtime; |
|
t[1].tv_sec = mtime; |
|
vfs_utimes(&wfile->f_path, t); |
|
|
|
fput(wfile); |
|
eat(body_len); |
|
state = SkipIt; |
|
return 0; |
|
} else { |
|
if (xwrite(wfile, victim, byte_count, &wfile_pos) != byte_count) |
|
error("write error"); |
|
body_len -= byte_count; |
|
eat(byte_count); |
|
return 1; |
|
} |
|
} |
|
|
|
static int __init do_symlink(void) |
|
{ |
|
collected[N_ALIGN(name_len) + body_len] = '\0'; |
|
clean_path(collected, 0); |
|
init_symlink(collected + N_ALIGN(name_len), collected); |
|
init_chown(collected, uid, gid, AT_SYMLINK_NOFOLLOW); |
|
do_utime(collected, mtime); |
|
state = SkipIt; |
|
next_state = Reset; |
|
return 0; |
|
} |
|
|
|
static __initdata int (*actions[])(void) = { |
|
[Start] = do_start, |
|
[Collect] = do_collect, |
|
[GotHeader] = do_header, |
|
[SkipIt] = do_skip, |
|
[GotName] = do_name, |
|
[CopyFile] = do_copy, |
|
[GotSymlink] = do_symlink, |
|
[Reset] = do_reset, |
|
}; |
|
|
|
static long __init write_buffer(char *buf, unsigned long len) |
|
{ |
|
byte_count = len; |
|
victim = buf; |
|
|
|
while (!actions[state]()) |
|
; |
|
return len - byte_count; |
|
} |
|
|
|
static long __init flush_buffer(void *bufv, unsigned long len) |
|
{ |
|
char *buf = (char *) bufv; |
|
long written; |
|
long origLen = len; |
|
if (message) |
|
return -1; |
|
while ((written = write_buffer(buf, len)) < len && !message) { |
|
char c = buf[written]; |
|
if (c == '0') { |
|
buf += written; |
|
len -= written; |
|
state = Start; |
|
} else if (c == 0) { |
|
buf += written; |
|
len -= written; |
|
state = Reset; |
|
} else |
|
error("junk within compressed archive"); |
|
} |
|
return origLen; |
|
} |
|
|
|
static unsigned long my_inptr; /* index of next byte to be processed in inbuf */ |
|
|
|
#include <linux/decompress/generic.h> |
|
|
|
static char * __init unpack_to_rootfs(char *buf, unsigned long len) |
|
{ |
|
long written; |
|
decompress_fn decompress; |
|
const char *compress_name; |
|
static __initdata char msg_buf[64]; |
|
|
|
header_buf = kmalloc(110, GFP_KERNEL); |
|
symlink_buf = kmalloc(PATH_MAX + N_ALIGN(PATH_MAX) + 1, GFP_KERNEL); |
|
name_buf = kmalloc(N_ALIGN(PATH_MAX), GFP_KERNEL); |
|
|
|
if (!header_buf || !symlink_buf || !name_buf) |
|
panic_show_mem("can't allocate buffers"); |
|
|
|
state = Start; |
|
this_header = 0; |
|
message = NULL; |
|
while (!message && len) { |
|
loff_t saved_offset = this_header; |
|
if (*buf == '0' && !(this_header & 3)) { |
|
state = Start; |
|
written = write_buffer(buf, len); |
|
buf += written; |
|
len -= written; |
|
continue; |
|
} |
|
if (!*buf) { |
|
buf++; |
|
len--; |
|
this_header++; |
|
continue; |
|
} |
|
this_header = 0; |
|
decompress = decompress_method(buf, len, &compress_name); |
|
pr_debug("Detected %s compressed data\n", compress_name); |
|
if (decompress) { |
|
int res = decompress(buf, len, NULL, flush_buffer, NULL, |
|
&my_inptr, error); |
|
if (res) |
|
error("decompressor failed"); |
|
} else if (compress_name) { |
|
if (!message) { |
|
snprintf(msg_buf, sizeof msg_buf, |
|
"compression method %s not configured", |
|
compress_name); |
|
message = msg_buf; |
|
} |
|
} else |
|
error("invalid magic at start of compressed archive"); |
|
if (state != Reset) |
|
error("junk at the end of compressed archive"); |
|
this_header = saved_offset + my_inptr; |
|
buf += my_inptr; |
|
len -= my_inptr; |
|
} |
|
dir_utime(); |
|
kfree(name_buf); |
|
kfree(symlink_buf); |
|
kfree(header_buf); |
|
return message; |
|
} |
|
|
|
static int __initdata do_retain_initrd; |
|
|
|
static int __init retain_initrd_param(char *str) |
|
{ |
|
if (*str) |
|
return 0; |
|
do_retain_initrd = 1; |
|
return 1; |
|
} |
|
__setup("retain_initrd", retain_initrd_param); |
|
|
|
#ifdef CONFIG_ARCH_HAS_KEEPINITRD |
|
static int __init keepinitrd_setup(char *__unused) |
|
{ |
|
do_retain_initrd = 1; |
|
return 1; |
|
} |
|
__setup("keepinitrd", keepinitrd_setup); |
|
#endif |
|
|
|
static bool __initdata initramfs_async = true; |
|
static int __init initramfs_async_setup(char *str) |
|
{ |
|
strtobool(str, &initramfs_async); |
|
return 1; |
|
} |
|
__setup("initramfs_async=", initramfs_async_setup); |
|
|
|
extern char __initramfs_start[]; |
|
extern unsigned long __initramfs_size; |
|
#include <linux/initrd.h> |
|
#include <linux/kexec.h> |
|
|
|
void __init reserve_initrd_mem(void) |
|
{ |
|
phys_addr_t start; |
|
unsigned long size; |
|
|
|
/* Ignore the virtul address computed during device tree parsing */ |
|
initrd_start = initrd_end = 0; |
|
|
|
if (!phys_initrd_size) |
|
return; |
|
/* |
|
* Round the memory region to page boundaries as per free_initrd_mem() |
|
* This allows us to detect whether the pages overlapping the initrd |
|
* are in use, but more importantly, reserves the entire set of pages |
|
* as we don't want these pages allocated for other purposes. |
|
*/ |
|
start = round_down(phys_initrd_start, PAGE_SIZE); |
|
size = phys_initrd_size + (phys_initrd_start - start); |
|
size = round_up(size, PAGE_SIZE); |
|
|
|
if (!memblock_is_region_memory(start, size)) { |
|
pr_err("INITRD: 0x%08llx+0x%08lx is not a memory region", |
|
(u64)start, size); |
|
goto disable; |
|
} |
|
|
|
if (memblock_is_region_reserved(start, size)) { |
|
pr_err("INITRD: 0x%08llx+0x%08lx overlaps in-use memory region\n", |
|
(u64)start, size); |
|
goto disable; |
|
} |
|
|
|
memblock_reserve(start, size); |
|
/* Now convert initrd to virtual addresses */ |
|
initrd_start = (unsigned long)__va(phys_initrd_start); |
|
initrd_end = initrd_start + phys_initrd_size; |
|
initrd_below_start_ok = 1; |
|
|
|
return; |
|
disable: |
|
pr_cont(" - disabling initrd\n"); |
|
initrd_start = 0; |
|
initrd_end = 0; |
|
} |
|
|
|
void __weak __init free_initrd_mem(unsigned long start, unsigned long end) |
|
{ |
|
#ifdef CONFIG_ARCH_KEEP_MEMBLOCK |
|
unsigned long aligned_start = ALIGN_DOWN(start, PAGE_SIZE); |
|
unsigned long aligned_end = ALIGN(end, PAGE_SIZE); |
|
|
|
memblock_free(__pa(aligned_start), aligned_end - aligned_start); |
|
#endif |
|
|
|
free_reserved_area((void *)start, (void *)end, POISON_FREE_INITMEM, |
|
"initrd"); |
|
} |
|
|
|
#ifdef CONFIG_KEXEC_CORE |
|
static bool __init kexec_free_initrd(void) |
|
{ |
|
unsigned long crashk_start = (unsigned long)__va(crashk_res.start); |
|
unsigned long crashk_end = (unsigned long)__va(crashk_res.end); |
|
|
|
/* |
|
* If the initrd region is overlapped with crashkernel reserved region, |
|
* free only memory that is not part of crashkernel region. |
|
*/ |
|
if (initrd_start >= crashk_end || initrd_end <= crashk_start) |
|
return false; |
|
|
|
/* |
|
* Initialize initrd memory region since the kexec boot does not do. |
|
*/ |
|
memset((void *)initrd_start, 0, initrd_end - initrd_start); |
|
if (initrd_start < crashk_start) |
|
free_initrd_mem(initrd_start, crashk_start); |
|
if (initrd_end > crashk_end) |
|
free_initrd_mem(crashk_end, initrd_end); |
|
return true; |
|
} |
|
#else |
|
static inline bool kexec_free_initrd(void) |
|
{ |
|
return false; |
|
} |
|
#endif /* CONFIG_KEXEC_CORE */ |
|
|
|
#ifdef CONFIG_BLK_DEV_RAM |
|
static void __init populate_initrd_image(char *err) |
|
{ |
|
ssize_t written; |
|
struct file *file; |
|
loff_t pos = 0; |
|
|
|
unpack_to_rootfs(__initramfs_start, __initramfs_size); |
|
|
|
printk(KERN_INFO "rootfs image is not initramfs (%s); looks like an initrd\n", |
|
err); |
|
file = filp_open("/initrd.image", O_WRONLY | O_CREAT, 0700); |
|
if (IS_ERR(file)) |
|
return; |
|
|
|
written = xwrite(file, (char *)initrd_start, initrd_end - initrd_start, |
|
&pos); |
|
if (written != initrd_end - initrd_start) |
|
pr_err("/initrd.image: incomplete write (%zd != %ld)\n", |
|
written, initrd_end - initrd_start); |
|
fput(file); |
|
} |
|
#endif /* CONFIG_BLK_DEV_RAM */ |
|
|
|
static void __init do_populate_rootfs(void *unused, async_cookie_t cookie) |
|
{ |
|
/* Load the built in initramfs */ |
|
char *err = unpack_to_rootfs(__initramfs_start, __initramfs_size); |
|
if (err) |
|
panic_show_mem("%s", err); /* Failed to decompress INTERNAL initramfs */ |
|
|
|
if (!initrd_start || IS_ENABLED(CONFIG_INITRAMFS_FORCE)) |
|
goto done; |
|
|
|
if (IS_ENABLED(CONFIG_BLK_DEV_RAM)) |
|
printk(KERN_INFO "Trying to unpack rootfs image as initramfs...\n"); |
|
else |
|
printk(KERN_INFO "Unpacking initramfs...\n"); |
|
|
|
err = unpack_to_rootfs((char *)initrd_start, initrd_end - initrd_start); |
|
if (err) { |
|
#ifdef CONFIG_BLK_DEV_RAM |
|
populate_initrd_image(err); |
|
#else |
|
printk(KERN_EMERG "Initramfs unpacking failed: %s\n", err); |
|
#endif |
|
} |
|
|
|
done: |
|
/* |
|
* If the initrd region is overlapped with crashkernel reserved region, |
|
* free only memory that is not part of crashkernel region. |
|
*/ |
|
if (!do_retain_initrd && initrd_start && !kexec_free_initrd()) |
|
free_initrd_mem(initrd_start, initrd_end); |
|
initrd_start = 0; |
|
initrd_end = 0; |
|
|
|
flush_delayed_fput(); |
|
} |
|
|
|
static ASYNC_DOMAIN_EXCLUSIVE(initramfs_domain); |
|
static async_cookie_t initramfs_cookie; |
|
|
|
void wait_for_initramfs(void) |
|
{ |
|
if (!initramfs_cookie) { |
|
/* |
|
* Something before rootfs_initcall wants to access |
|
* the filesystem/initramfs. Probably a bug. Make a |
|
* note, avoid deadlocking the machine, and let the |
|
* caller's access fail as it used to. |
|
*/ |
|
pr_warn_once("wait_for_initramfs() called before rootfs_initcalls\n"); |
|
return; |
|
} |
|
async_synchronize_cookie_domain(initramfs_cookie + 1, &initramfs_domain); |
|
} |
|
EXPORT_SYMBOL_GPL(wait_for_initramfs); |
|
|
|
static int __init populate_rootfs(void) |
|
{ |
|
initramfs_cookie = async_schedule_domain(do_populate_rootfs, NULL, |
|
&initramfs_domain); |
|
if (!initramfs_async) |
|
wait_for_initramfs(); |
|
return 0; |
|
} |
|
rootfs_initcall(populate_rootfs);
|
|
|