Commit 1438f013 authored by Christopher Covington's avatar Christopher Covington Committed by Pavel Emelyanov

Pass task_size to vma_area_is_private()

If we want one CRIU binary to work across all AArch64 kernel
configurations, a single task size value cannot be hard coded. Since
vma_area_is_private() is used by both restorer blob code and non
restorer blob code, which must use different variables for recording
the task size, make task_size a function argument and modify the call
sites accordingly. This fixes the following error on AArch64 kernels
with CONFIG_ARM64_64K_PAGES=y.

  pie: Error (pie/restorer.c:929): Can't restore 0x3ffb7e70000 mapping w>
  pie: ith 0xfffffffffffffff7
Signed-off-by: 's avatarChristopher Covington <cov@codeaurora.org>
Signed-off-by: 's avatarPavel Emelyanov <xemul@parallels.com>
parent 7451fc7d
...@@ -269,7 +269,7 @@ static int map_private_vma(struct vma_area *vma, void **tgt_addr, ...@@ -269,7 +269,7 @@ static int map_private_vma(struct vma_area *vma, void **tgt_addr,
if (p->e->start > vma->e->start) if (p->e->start > vma->e->start)
break; break;
if (!vma_area_is_private(p)) if (!vma_area_is_private(p, kdat.task_size))
continue; continue;
if (p->e->end != vma->e->end || if (p->e->end != vma->e->end ||
...@@ -381,7 +381,7 @@ static int premap_priv_vmas(struct vm_area_list *vmas, void *at) ...@@ -381,7 +381,7 @@ static int premap_priv_vmas(struct vm_area_list *vmas, void *at)
} }
pstart = vma->e->start; pstart = vma->e->start;
if (!vma_area_is_private(vma)) if (!vma_area_is_private(vma, kdat.task_size))
continue; continue;
ret = map_private_vma(vma, &at, &pvma, parent_vmas); ret = map_private_vma(vma, &at, &pvma, parent_vmas);
...@@ -447,7 +447,7 @@ static int restore_priv_vma_content(void) ...@@ -447,7 +447,7 @@ static int restore_priv_vma_content(void)
*/ */
if (va < vma->e->start) if (va < vma->e->start)
goto err_addr; goto err_addr;
else if (unlikely(!vma_area_is_private(vma))) { else if (unlikely(!vma_area_is_private(vma, kdat.task_size))) {
pr_err("Trying to restore page for non-private VMA\n"); pr_err("Trying to restore page for non-private VMA\n");
goto err_addr; goto err_addr;
} }
...@@ -590,7 +590,7 @@ static int unmap_guard_pages() ...@@ -590,7 +590,7 @@ static int unmap_guard_pages()
struct list_head *vmas = &rsti(current)->vmas.h; struct list_head *vmas = &rsti(current)->vmas.h;
list_for_each_entry(vma, vmas, list) { list_for_each_entry(vma, vmas, list) {
if (!vma_area_is_private(vma)) if (!vma_area_is_private(vma, kdat.task_size))
continue; continue;
if (vma->e->flags & MAP_GROWSDOWN) { if (vma->e->flags & MAP_GROWSDOWN) {
...@@ -2713,7 +2713,7 @@ static int sigreturn_restore(pid_t pid, CoreEntry *core) ...@@ -2713,7 +2713,7 @@ static int sigreturn_restore(pid_t pid, CoreEntry *core)
*vme = *vma->e; *vme = *vma->e;
if (vma_area_is_private(vma)) if (vma_area_is_private(vma, kdat.task_size))
vma_premmaped_start(vme) = vma->premmaped_addr; vma_premmaped_start(vme) = vma->premmaped_addr;
} }
......
...@@ -92,17 +92,19 @@ static inline int in_vma_area(struct vma_area *vma, unsigned long addr) ...@@ -92,17 +92,19 @@ static inline int in_vma_area(struct vma_area *vma, unsigned long addr)
addr < (unsigned long)vma->e->end; addr < (unsigned long)vma->e->end;
} }
static inline bool vma_entry_is_private(VmaEntry *entry) static inline bool vma_entry_is_private(VmaEntry *entry,
unsigned long task_size)
{ {
return vma_entry_is(entry, VMA_AREA_REGULAR) && return vma_entry_is(entry, VMA_AREA_REGULAR) &&
(vma_entry_is(entry, VMA_ANON_PRIVATE) || (vma_entry_is(entry, VMA_ANON_PRIVATE) ||
vma_entry_is(entry, VMA_FILE_PRIVATE)) && vma_entry_is(entry, VMA_FILE_PRIVATE)) &&
(entry->end <= TASK_SIZE); (entry->end <= task_size);
} }
static inline bool vma_area_is_private(struct vma_area *vma) static inline bool vma_area_is_private(struct vma_area *vma,
unsigned long task_size)
{ {
return vma_entry_is_private(vma->e); return vma_entry_is_private(vma->e, task_size);
} }
#endif /* __CR_VMA_H__ */ #endif /* __CR_VMA_H__ */
...@@ -178,7 +178,7 @@ static struct parasite_dump_pages_args *prep_dump_pages_args(struct parasite_ctl ...@@ -178,7 +178,7 @@ static struct parasite_dump_pages_args *prep_dump_pages_args(struct parasite_ctl
args->nr_vmas = 0; args->nr_vmas = 0;
list_for_each_entry(vma, &vma_area_list->h, list) { list_for_each_entry(vma, &vma_area_list->h, list) {
if (!vma_area_is_private(vma)) if (!vma_area_is_private(vma, kdat.task_size))
continue; continue;
if (vma->e->prot & PROT_READ) if (vma->e->prot & PROT_READ)
continue; continue;
...@@ -293,7 +293,7 @@ static int __parasite_dump_pages_seized(struct parasite_ctl *ctl, ...@@ -293,7 +293,7 @@ static int __parasite_dump_pages_seized(struct parasite_ctl *ctl,
u64 off = 0; u64 off = 0;
u64 *map; u64 *map;
if (!vma_area_is_private(vma_area)) if (!vma_area_is_private(vma_area, kdat.task_size))
continue; continue;
map = pmc_get_map(&pmc, vma_area); map = pmc_get_map(&pmc, vma_area);
...@@ -441,7 +441,7 @@ int prepare_mm_pid(struct pstree_item *i) ...@@ -441,7 +441,7 @@ int prepare_mm_pid(struct pstree_item *i)
} }
list_add_tail(&vma->list, &ri->vmas.h); list_add_tail(&vma->list, &ri->vmas.h);
if (vma_area_is_private(vma)) { if (vma_area_is_private(vma, kdat.task_size)) {
ri->vmas.priv_size += vma_area_len(vma); ri->vmas.priv_size += vma_area_len(vma);
if (vma->e->flags & MAP_GROWSDOWN) if (vma->e->flags & MAP_GROWSDOWN)
ri->vmas.priv_size += PAGE_SIZE; ri->vmas.priv_size += PAGE_SIZE;
......
...@@ -878,7 +878,7 @@ long __export_restore_task(struct task_restore_args *args) ...@@ -878,7 +878,7 @@ long __export_restore_task(struct task_restore_args *args)
for (i = 0; i < args->vmas_n; i++) { for (i = 0; i < args->vmas_n; i++) {
vma_entry = args->vmas + i; vma_entry = args->vmas + i;
if (!vma_entry_is_private(vma_entry)) if (!vma_entry_is_private(vma_entry, args->task_size))
continue; continue;
if (vma_entry->end >= args->task_size) if (vma_entry->end >= args->task_size)
...@@ -896,7 +896,7 @@ long __export_restore_task(struct task_restore_args *args) ...@@ -896,7 +896,7 @@ long __export_restore_task(struct task_restore_args *args)
for (i = args->vmas_n - 1; i >= 0; i--) { for (i = args->vmas_n - 1; i >= 0; i--) {
vma_entry = args->vmas + i; vma_entry = args->vmas + i;
if (!vma_entry_is_private(vma_entry)) if (!vma_entry_is_private(vma_entry, args->task_size))
continue; continue;
if (vma_entry->start > args->task_size) if (vma_entry->start > args->task_size)
...@@ -919,7 +919,7 @@ long __export_restore_task(struct task_restore_args *args) ...@@ -919,7 +919,7 @@ long __export_restore_task(struct task_restore_args *args)
if (!vma_entry_is(vma_entry, VMA_AREA_REGULAR)) if (!vma_entry_is(vma_entry, VMA_AREA_REGULAR))
continue; continue;
if (vma_entry_is_private(vma_entry)) if (vma_entry_is_private(vma_entry, args->task_size))
continue; continue;
va = restore_mapping(vma_entry); va = restore_mapping(vma_entry);
......
...@@ -473,7 +473,7 @@ static int vma_list_add(struct vma_area *vma_area, ...@@ -473,7 +473,7 @@ static int vma_list_add(struct vma_area *vma_area,
list_add_tail(&vma_area->list, &vma_area_list->h); list_add_tail(&vma_area->list, &vma_area_list->h);
vma_area_list->nr++; vma_area_list->nr++;
if (vma_area_is_private(vma_area)) { if (vma_area_is_private(vma_area, kdat.task_size)) {
unsigned long pages; unsigned long pages;
pages = vma_area_len(vma_area) / PAGE_SIZE; pages = vma_area_len(vma_area) / PAGE_SIZE;
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment