*
* returns the time in ktime_t format
*/
-static ktime_t ktime_get(void)
+ktime_t ktime_get(void)
{
struct timespec now;
*
* returns the time in ktime_t format
*/
-static ktime_t ktime_get_real(void)
+ktime_t ktime_get_real(void)
{
struct timespec now;
*/
#ifdef CONFIG_SMP
-#define set_curr_timer(b, t) do { (b)->curr_timer = (t); } while (0)
-
/*
* We are using hashed locking: holding per_cpu(hrtimer_bases)[n].lock
* means that all timers which are tied to this base via timer->base are
* completed. There is no conflict as we hold the lock until
* the timer is enqueued.
*/
- if (unlikely(base->cpu_base->curr_timer == timer))
+ if (unlikely(timer->state & HRTIMER_STATE_CALLBACK))
return base;
/* See the comment in lock_timer_base() */
#else /* CONFIG_SMP */
-#define set_curr_timer(b, t) do { } while (0)
-
static inline struct hrtimer_clock_base *
lock_hrtimer_base(const struct hrtimer *timer, unsigned long *flags)
{
# define ktime_divns(kt, div) (unsigned long)((kt).tv64 / (div))
#endif /* BITS_PER_LONG >= 64 */
-/*
- * Timekeeping resumed notification
- */
-void hrtimer_notify_resume(void)
-{
- clock_was_set();
-}
-
/*
* Counterpart to lock_timer_base above:
*/
break;
fn = timer->function;
- set_curr_timer(cpu_base, timer);
__remove_hrtimer(timer, base, HRTIMER_STATE_CALLBACK);
spin_unlock_irq(&cpu_base->lock);
enqueue_hrtimer(timer, base);
}
}
- set_curr_timer(cpu_base, NULL);
spin_unlock_irq(&cpu_base->lock);
}
spin_lock(&old_base->lock);
for (i = 0; i < HRTIMER_MAX_CLOCK_BASES; i++) {
- BUG_ON(old_base->curr_timer);
-
migrate_hrtimer_list(&old_base->clock_base[i],
&new_base->clock_base[i]);
}
#ifdef CONFIG_HOTPLUG_CPU
case CPU_DEAD:
+ clockevents_notify(CLOCK_EVT_NOTIFY_CPU_DEAD, &cpu);
migrate_hrtimers(cpu);
break;
#endif