mirror of
https://github.com/adulau/aha.git
synced 2024-12-28 03:36:19 +00:00
time: fix typo in comments
Fix typo in comments. BTW: I have to fix coding style in arch/ia64/kernel/time.c also, otherwise checkpatch.pl will be complaining. Signed-off-by: Li Zefan <lizf@cn.fujitsu.com> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Ingo Molnar <mingo@elte.hu> Cc: john stultz <johnstul@us.ibm.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
cf4fc6cb76
commit
3eb056764d
7 changed files with 16 additions and 16 deletions
|
@ -53,7 +53,7 @@ static struct clocksource clocksource_itc = {
|
|||
.rating = 350,
|
||||
.read = itc_get_cycles,
|
||||
.mask = CLOCKSOURCE_MASK(64),
|
||||
.mult = 0, /*to be caluclated*/
|
||||
.mult = 0, /*to be calculated*/
|
||||
.shift = 16,
|
||||
.flags = CLOCK_SOURCE_IS_CONTINUOUS,
|
||||
};
|
||||
|
|
|
@ -77,7 +77,7 @@ unsigned long __init native_calculate_cpu_khz(void)
|
|||
reserve_evntsel_nmi(MSR_K7_EVNTSEL0 + i);
|
||||
}
|
||||
local_irq_save(flags);
|
||||
/* start meauring cycles, incrementing from 0 */
|
||||
/* start measuring cycles, incrementing from 0 */
|
||||
wrmsrl(MSR_K7_PERFCTR0 + i, 0);
|
||||
wrmsrl(MSR_K7_EVNTSEL0 + i, 1 << 22 | 3 << 16 | 0x76);
|
||||
rdtscl(tsc_start);
|
||||
|
|
|
@ -78,7 +78,7 @@ enum hrtimer_cb_mode {
|
|||
* as otherwise the timer could be removed before the softirq code finishes the
|
||||
* the handling of the timer.
|
||||
*
|
||||
* The HRTIMER_STATE_ENQUEUE bit is always or'ed to the current state to
|
||||
* The HRTIMER_STATE_ENQUEUED bit is always or'ed to the current state to
|
||||
* preserve the HRTIMER_STATE_CALLBACK bit in the above scenario.
|
||||
*
|
||||
* All state transitions are protected by cpu_base->lock.
|
||||
|
|
|
@ -42,7 +42,7 @@
|
|||
/* LATCH is used in the interval timer and ftape setup. */
|
||||
#define LATCH ((CLOCK_TICK_RATE + HZ/2) / HZ) /* For divider */
|
||||
|
||||
/* Suppose we want to devide two numbers NOM and DEN: NOM/DEN, the we can
|
||||
/* Suppose we want to devide two numbers NOM and DEN: NOM/DEN, then we can
|
||||
* improve accuracy by shifting LSH bits, hence calculating:
|
||||
* (NOM << LSH) / DEN
|
||||
* This however means trouble for large NOM, because (NOM << LSH) may no
|
||||
|
@ -160,7 +160,7 @@ extern unsigned long preset_lpj;
|
|||
* We want to do realistic conversions of time so we need to use the same
|
||||
* values the update wall clock code uses as the jiffies size. This value
|
||||
* is: TICK_NSEC (which is defined in timex.h). This
|
||||
* is a constant and is in nanoseconds. We will used scaled math
|
||||
* is a constant and is in nanoseconds. We will use scaled math
|
||||
* with a set of scales defined here as SEC_JIFFIE_SC, USEC_JIFFIE_SC and
|
||||
* NSEC_JIFFIE_SC. Note that these defines contain nothing but
|
||||
* constants and so are computed at compile time. SHIFT_HZ (computed in
|
||||
|
@ -204,7 +204,7 @@ extern unsigned long preset_lpj;
|
|||
* operator if the result is a long long AND at least one of the
|
||||
* operands is cast to long long (usually just prior to the "*" so as
|
||||
* not to confuse it into thinking it really has a 64-bit operand,
|
||||
* which, buy the way, it can do, but it take more code and at least 2
|
||||
* which, buy the way, it can do, but it takes more code and at least 2
|
||||
* mpys).
|
||||
|
||||
* We also need to be aware that one second in nanoseconds is only a
|
||||
|
|
|
@ -267,7 +267,7 @@ EXPORT_SYMBOL(jiffies_to_usecs);
|
|||
*
|
||||
* This function should be only used for timestamps returned by
|
||||
* current_kernel_time() or CURRENT_TIME, not with do_gettimeofday() because
|
||||
* it doesn't handle the better resolution of the later.
|
||||
* it doesn't handle the better resolution of the latter.
|
||||
*/
|
||||
struct timespec timespec_trunc(struct timespec t, unsigned gran)
|
||||
{
|
||||
|
@ -315,7 +315,7 @@ EXPORT_SYMBOL_GPL(getnstimeofday);
|
|||
* This algorithm was first published by Gauss (I think).
|
||||
*
|
||||
* WARNING: this function will overflow on 2106-02-07 06:28:16 on
|
||||
* machines were long is 32-bit! (However, as time_t is signed, we
|
||||
* machines where long is 32-bit! (However, as time_t is signed, we
|
||||
* will already get problems at other places on 2038-01-19 03:14:08)
|
||||
*/
|
||||
unsigned long
|
||||
|
|
|
@ -133,7 +133,7 @@ static void clockevents_do_notify(unsigned long reason, void *dev)
|
|||
}
|
||||
|
||||
/*
|
||||
* Called after a notify add to make devices availble which were
|
||||
* Called after a notify add to make devices available which were
|
||||
* released from the notifier call.
|
||||
*/
|
||||
static void clockevents_notify_released(void)
|
||||
|
|
|
@ -364,7 +364,7 @@ static __always_inline int clocksource_bigadjust(s64 error, s64 *interval,
|
|||
* with losing too many ticks, otherwise we would overadjust and
|
||||
* produce an even larger error. The smaller the adjustment the
|
||||
* faster we try to adjust for it, as lost ticks can do less harm
|
||||
* here. This is tuned so that an error of about 1 msec is adusted
|
||||
* here. This is tuned so that an error of about 1 msec is adjusted
|
||||
* within about 1 sec (or 2^20 nsec in 2^SHIFT_HZ ticks).
|
||||
*/
|
||||
error2 = clock->error >> (TICK_LENGTH_SHIFT + 22 - 2 * SHIFT_HZ);
|
||||
|
|
Loading…
Reference in a new issue