eal/armv8: fix timer frequency calibration with PMU
authorHonnappa Nagarahalli <honnappa.nagarahalli@arm.com>
Fri, 26 Jun 2020 20:35:01 +0000 (15:35 -0500)
committerDavid Marchand <david.marchand@redhat.com>
Tue, 7 Jul 2020 11:20:50 +0000 (13:20 +0200)
get_tsc_freq uses 'nanosleep' system call to calculate the CPU
frequency. However, 'nanosleep' results in the process getting
un-scheduled. The kernel saves and restores the PMU state. This
ensures that the PMU cycles are not counted towards a sleeping
process. When RTE_ARM_EAL_RDTSC_USE_PMU is defined, this results
in incorrect CPU frequency calculation. This logic is replaced
with generic counter based loop.

Bugzilla ID: 450
Fixes: f91bcbb2d9a6 ("eal/armv8: use high-resolution cycle counter")
Cc: stable@dpdk.org
Signed-off-by: Honnappa Nagarahalli <honnappa.nagarahalli@arm.com>
Reviewed-by: Ruifeng Wang <ruifeng.wang@arm.com>
Reviewed-by: Dharmik Thakkar <dharmik.thakkar@arm.com>
Reviewed-by: Phil Yang <phil.yang@arm.com>
Acked-by: Jerin Jacob <jerinj@marvell.com>
lib/librte_eal/arm/include/rte_cycles_64.h
lib/librte_eal/arm/rte_cycles.c

index da557b6..e41f9db 100644 (file)
@@ -1,5 +1,6 @@
 /* SPDX-License-Identifier: BSD-3-Clause
  * Copyright(c) 2015 Cavium, Inc
+ * Copyright(c) 2020 Arm Limited
  */
 
 #ifndef _RTE_CYCLES_ARM64_H_
@@ -11,6 +12,33 @@ extern "C" {
 
 #include "generic/rte_cycles.h"
 
+/** Read generic counter frequency */
+static __rte_always_inline uint64_t
+__rte_arm64_cntfrq(void)
+{
+       uint64_t freq;
+
+       asm volatile("mrs %0, cntfrq_el0" : "=r" (freq));
+       return freq;
+}
+
+/** Read generic counter */
+static __rte_always_inline uint64_t
+__rte_arm64_cntvct(void)
+{
+       uint64_t tsc;
+
+       asm volatile("mrs %0, cntvct_el0" : "=r" (tsc));
+       return tsc;
+}
+
+static __rte_always_inline uint64_t
+__rte_arm64_cntvct_precise(void)
+{
+       asm volatile("isb" : : : "memory");
+       return __rte_arm64_cntvct();
+}
+
 /**
  * Read the time base register.
  *
@@ -25,10 +53,7 @@ extern "C" {
 static inline uint64_t
 rte_rdtsc(void)
 {
-       uint64_t tsc;
-
-       asm volatile("mrs %0, cntvct_el0" : "=r" (tsc));
-       return tsc;
+       return __rte_arm64_cntvct();
 }
 #else
 /**
@@ -49,14 +74,22 @@ rte_rdtsc(void)
  * asm volatile("msr pmcr_el0, %0" : : "r" (val));
  *
  */
-static inline uint64_t
-rte_rdtsc(void)
+
+/** Read PMU cycle counter */
+static __rte_always_inline uint64_t
+__rte_arm64_pmccntr(void)
 {
        uint64_t tsc;
 
        asm volatile("mrs %0, pmccntr_el0" : "=r"(tsc));
        return tsc;
 }
+
+static inline uint64_t
+rte_rdtsc(void)
+{
+       return __rte_arm64_pmccntr();
+}
 #endif
 
 static inline uint64_t
index 3500d52..5bd29b2 100644 (file)
@@ -3,14 +3,35 @@
  */
 
 #include "eal_private.h"
+#include "rte_cycles.h"
 
 uint64_t
 get_tsc_freq_arch(void)
 {
 #if defined RTE_ARCH_ARM64 && !defined RTE_ARM_EAL_RDTSC_USE_PMU
-       uint64_t freq;
-       asm volatile("mrs %0, cntfrq_el0" : "=r" (freq));
-       return freq;
+       return __rte_arm64_cntfrq();
+#elif defined RTE_ARCH_ARM64 && defined RTE_ARM_EAL_RDTSC_USE_PMU
+#define CYC_PER_1MHZ 1E6
+       /* Use the generic counter ticks to calculate the PMU
+        * cycle frequency.
+        */
+       uint64_t ticks;
+       uint64_t start_ticks, cur_ticks;
+       uint64_t start_pmu_cycles, end_pmu_cycles;
+
+       /* Number of ticks for 1/10 second */
+       ticks = __rte_arm64_cntfrq() / 10;
+
+       start_ticks = __rte_arm64_cntvct_precise();
+       start_pmu_cycles = rte_rdtsc_precise();
+       do {
+               cur_ticks = __rte_arm64_cntvct();
+       } while ((cur_ticks - start_ticks) < ticks);
+       end_pmu_cycles = rte_rdtsc_precise();
+
+       /* Adjust the cycles to next 1Mhz */
+       return RTE_ALIGN_MUL_CEIL(end_pmu_cycles - start_pmu_cycles,
+                       CYC_PER_1MHZ) * 10;
 #else
        return 0;
 #endif