remove unneeded tests for NULL when freeing
[dpdk.git] / lib / librte_eal / linuxapp / eal / eal_memory.c
index ac2745e..6008533 100644 (file)
 #include "eal_filesystem.h"
 #include "eal_hugepages.h"
 
+#ifdef RTE_LIBRTE_XEN_DOM0
+int rte_xen_dom0_supported(void)
+{
+       return internal_config.xen_dom0_support;
+}
+#endif
+
 /**
  * @file
  * Huge page mapping under linux
@@ -392,8 +399,10 @@ map_all_hugepages(struct hugepage_file *hugepg_tbl,
                        return -1;
                }
 
+               /* map the segment, and populate page tables,
+                * the kernel fills this segment with zeros */
                virtaddr = mmap(vma_addr, hugepage_sz, PROT_READ | PROT_WRITE,
-                               MAP_SHARED, fd, 0);
+                               MAP_SHARED | MAP_POPULATE, fd, 0);
                if (virtaddr == MAP_FAILED) {
                        RTE_LOG(ERR, EAL, "%s(): mmap failed: %s\n", __func__,
                                        strerror(errno));
@@ -403,7 +412,6 @@ map_all_hugepages(struct hugepage_file *hugepg_tbl,
 
                if (orig) {
                        hugepg_tbl[i].orig_va = virtaddr;
-                       memset(virtaddr, 0, hugepage_sz);
                }
                else {
                        hugepg_tbl[i].final_va = virtaddr;
@@ -522,22 +530,16 @@ remap_all_hugepages(struct hugepage_file *hugepg_tbl, struct hugepage_info *hpi)
 
                        old_addr = vma_addr;
 
-                       /* map new, bigger segment */
+                       /* map new, bigger segment, and populate page tables,
+                        * the kernel fills this segment with zeros */
                        vma_addr = mmap(vma_addr, total_size,
-                                       PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
+                                       PROT_READ | PROT_WRITE, MAP_SHARED | MAP_POPULATE, fd, 0);
 
                        if (vma_addr == MAP_FAILED || vma_addr != old_addr) {
                                RTE_LOG(ERR, EAL, "%s(): mmap failed: %s\n", __func__, strerror(errno));
                                close(fd);
                                return -1;
                        }
-
-                       /* touch the page. this is needed because kernel postpones mapping
-                        * creation until the first page fault. with this, we pin down
-                        * the page and it is marked as used and gets into process' pagemap.
-                        */
-                       for (offset = 0; offset < total_size; offset += hugepage_sz)
-                               *((volatile uint8_t*) RTE_PTR_ADD(vma_addr, offset));
                }
 
                /* set shared flock on the file. */
@@ -585,9 +587,6 @@ remap_all_hugepages(struct hugepage_file *hugepg_tbl, struct hugepage_info *hpi)
                        }
                }
 
-               /* zero out the whole segment */
-               memset(hugepg_tbl[page_idx].final_va, 0, total_size);
-
                page_idx++;
        }
 
@@ -786,6 +785,30 @@ copy_hugepages_to_shared_mem(struct hugepage_file * dst, int dest_size,
        return 0;
 }
 
+static int
+unlink_hugepage_files(struct hugepage_file *hugepg_tbl,
+               unsigned num_hp_info)
+{
+       unsigned socket, size;
+       int page, nrpages = 0;
+
+       /* get total number of hugepages */
+       for (size = 0; size < num_hp_info; size++)
+               for (socket = 0; socket < RTE_MAX_NUMA_NODES; socket++)
+                       nrpages +=
+                       internal_config.hugepage_info[size].num_pages[socket];
+
+       for (page = 0; page < nrpages; page++) {
+               struct hugepage_file *hp = &hugepg_tbl[page];
+
+               if (hp->final_va != NULL && unlink(hp->filepath)) {
+                       RTE_LOG(WARNING, EAL, "%s(): Removing %s failed: %s\n",
+                               __func__, hp->filepath, strerror(errno));
+               }
+       }
+       return 0;
+}
+
 /*
  * unmaps hugepages that are not going to be used. since we originally allocate
  * ALL hugepages (not just those we need), additional unmapping needs to be done.
@@ -1289,6 +1312,13 @@ rte_eal_hugepage_init(void)
                goto fail;
        }
 
+       /* free the hugepage backing files */
+       if (internal_config.hugepage_unlink &&
+               unlink_hugepage_files(tmp_hp, internal_config.num_hugepage_sizes) < 0) {
+               RTE_LOG(ERR, EAL, "Unlinking hugepage files failed!\n");
+               goto fail;
+       }
+
        /* free the temporary hugepage table */
        free(tmp_hp);
        tmp_hp = NULL;
@@ -1374,8 +1404,7 @@ rte_eal_hugepage_init(void)
        return 0;
 
 fail:
-       if (tmp_hp)
-               free(tmp_hp);
+       free(tmp_hp);
        return -1;
 }