* operations into the same table.
*
* The typical reason an implementation may choose to split the table lookup
- * operation into multiple steps is to hide the latency of the inherrent memory
+ * operation into multiple steps is to hide the latency of the inherent memory
* read operations: before a read operation with the source data likely not in
* the CPU cache, the source data prefetch is issued and the table lookup
* operation is postponed in favor of some other unrelated work, which the CPU