From 69f1678b4bf3d7da857766e196bc1449e6b78c2a Mon Sep 17 00:00:00 2001 From: Bryan Cantrill Date: Fri, 13 May 2011 22:54:50 -0700 Subject: HVM-201 all PIT functionality should be in kvm_i8254.c --- kvm_i8254.c | 720 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++--- 1 file changed, 694 insertions(+), 26 deletions(-) (limited to 'kvm_i8254.c') diff --git a/kvm_i8254.c b/kvm_i8254.c index b7b890d..23708fd 100644 --- a/kvm_i8254.c +++ b/kvm_i8254.c @@ -45,6 +45,197 @@ extern int kvm_vcpu_is_bsp(struct kvm_vcpu *vcpu); +static uint64_t +muldiv64(uint64_t a, uint32_t b, uint32_t c) +{ + union { + uint64_t ll; + struct { + uint32_t low, high; + } l; + } u, res; + uint64_t rl, rh; + + u.ll = a; + rl = (uint64_t)u.l.low * (uint64_t)b; + rh = (uint64_t)u.l.high * (uint64_t)b; + rh += (rl >> 32); + res.l.high = rh/c; + res.l.low = ((mod_64(rh, c) << 32) + (rl & 0xffffffff))/ c; + + return (res.ll); +} + +static void +pit_set_gate(struct kvm *kvm, int channel, uint32_t val) +{ + struct kvm_kpit_channel_state *c = + &kvm->arch.vpit->pit_state.channels[channel]; + + ASSERT(mutex_owned(&kvm->arch.vpit->pit_state.lock)); + + switch (c->mode) { + default: + case 0: + case 4: + /* XXX: just disable/enable counting */ + break; + case 1: + case 2: + case 3: + case 5: + /* Restart counting on rising edge. */ + if (c->gate < val) + c->count_load_time = gethrtime(); + break; + } + + c->gate = val; +} + +static int +pit_get_gate(struct kvm *kvm, int channel) +{ + ASSERT(mutex_owned(&kvm->arch.vpit->pit_state.lock)); + + return (kvm->arch.vpit->pit_state.channels[channel].gate); +} + +static int64_t +__kpit_elapsed(struct kvm *kvm) +{ + int64_t elapsed; + hrtime_t remaining, now; + struct kvm_kpit_state *ps = &kvm->arch.vpit->pit_state; + + if (!ps->pit_timer.period) + return (0); + + /* + * The Counter does not stop when it reaches zero. In + * Modes 0, 1, 4, and 5 the Counter ``wraps around'' to + * the highest count, either FFFF hex for binary counting + * or 9999 for BCD counting, and continues counting. + * Modes 2 and 3 are periodic; the Counter reloads + * itself with the initial count and continues counting + * from there. + */ + now = gethrtime(); + elapsed = now - ps->pit_timer.start - + ps->pit_timer.period * ps->pit_timer.intervals; + remaining = ps->pit_timer.period - elapsed; + elapsed = mod_64(elapsed, ps->pit_timer.period); + + return (elapsed); +} + +static int64_t +kpit_elapsed(struct kvm *kvm, struct kvm_kpit_channel_state *c, int channel) +{ + if (channel == 0) + return (__kpit_elapsed(kvm)); + + return (gethrtime() - c->count_load_time); +} + +static int +pit_get_count(struct kvm *kvm, int channel) +{ + struct kvm_kpit_channel_state *c = + &kvm->arch.vpit->pit_state.channels[channel]; + int64_t d, t; + int counter; + + ASSERT(mutex_owned(&kvm->arch.vpit->pit_state.lock)); + + t = kpit_elapsed(kvm, c, channel); + d = muldiv64(t, KVM_PIT_FREQ, NSEC_PER_SEC); + + switch (c->mode) { + case 0: + case 1: + case 4: + case 5: + counter = (c->count - d) & 0xffff; + break; + case 3: + /* XXX: may be incorrect for odd counts */ + counter = c->count - (mod_64((2 * d), c->count)); + break; + default: + counter = c->count - mod_64(d, c->count); + break; + } + + return (counter); +} + +static int +pit_get_out(struct kvm *kvm, int channel) +{ + struct kvm_kpit_channel_state *c = + &kvm->arch.vpit->pit_state.channels[channel]; + int64_t d, t; + int out; + + ASSERT(mutex_owned(&kvm->arch.vpit->pit_state.lock)); + + t = kpit_elapsed(kvm, c, channel); + d = muldiv64(t, KVM_PIT_FREQ, NSEC_PER_SEC); + + switch (c->mode) { + default: + case 0: + out = (d >= c->count); + break; + case 1: + out = (d < c->count); + break; + case 2: + out = ((mod_64(d, c->count) == 0) && (d != 0)); + break; + case 3: + out = (mod_64(d, c->count) < ((c->count + 1) >> 1)); + break; + case 4: + case 5: + out = (d == c->count); + break; + } + + return (out); +} + +static void +pit_latch_count(struct kvm *kvm, int channel) +{ + struct kvm_kpit_channel_state *c = + &kvm->arch.vpit->pit_state.channels[channel]; + + ASSERT(mutex_owned(&kvm->arch.vpit->pit_state.lock)); + + if (!c->count_latched) { + c->latched_count = pit_get_count(kvm, channel); + c->count_latched = c->rw_mode; + } +} + +static void +pit_latch_status(struct kvm *kvm, int channel) +{ + struct kvm_kpit_channel_state *c = + &kvm->arch.vpit->pit_state.channels[channel]; + + ASSERT(mutex_owned(&kvm->arch.vpit->pit_state.lock)); + + if (!c->status_latched) { + /* TODO: Return NULL COUNT (bit 6). */ + c->status = ((pit_get_out(kvm, channel) << 7) | + (c->rw_mode << 4) | (c->mode << 1) | c->bcd); + c->status_latched = 1; + } +} + int pit_has_pending_timer(struct kvm_vcpu *vcpu) { @@ -61,6 +252,509 @@ pit_has_pending_timer(struct kvm_vcpu *vcpu) return (0); } +static void +kvm_pit_ack_irq(struct kvm_irq_ack_notifier *kian) +{ + struct kvm_kpit_state *ps = (struct kvm_kpit_state *)(((caddr_t)kian) - + offsetof(struct kvm_kpit_state, irq_ack_notifier)); + mutex_enter(&ps->inject_lock); + if (--ps->pit_timer.pending < 0) + ps->pit_timer.pending++; + ps->irq_ack = 1; + mutex_exit(&ps->inject_lock); +} + +static void +destroy_pit_timer(struct kvm_timer *pt) +{ +#ifdef XXX + pr_debug("pit: " "execute del timer!\n"); + hrtimer_cancel_p(&pt->timer); +#else + XXX_KVM_PROBE; +#endif +} + +static int +kpit_is_periodic(struct kvm_timer *ktimer) +{ + struct kvm_kpit_state *ps = (struct kvm_kpit_state *) + (((caddr_t)ktimer) - offsetof(struct kvm_kpit_state, pit_timer)); + + return (ps->is_periodic); +} + +static struct kvm_timer_ops kpit_ops = { + .is_periodic = kpit_is_periodic, +}; + +static void +create_pit_timer(struct kvm_kpit_state *ps, uint32_t val, int is_period) +{ + struct kvm_timer *pt = &ps->pit_timer; + int64_t interval; + + interval = muldiv64(val, NSEC_PER_SEC, KVM_PIT_FREQ); + + mutex_enter(&cpu_lock); + + /* TODO The new value only affected after the retriggered */ + if (pt->active) { + cyclic_remove(pt->kvm_cyclic_id); + pt->active = 0; + } + pt->period = interval; + ps->is_periodic = is_period; + + pt->kvm_cyc_handler.cyh_func = kvm_timer_fire; + pt->kvm_cyc_handler.cyh_level = CY_LOW_LEVEL; + pt->kvm_cyc_handler.cyh_arg = pt; + pt->t_ops = &kpit_ops; + pt->kvm = ps->pit->kvm; + pt->vcpu = pt->kvm->bsp_vcpu; + + pt->pending = 0; /* XXX need protection? */ + ps->irq_ack = 1; + pt->start = gethrtime(); + + if (is_period) { + pt->kvm_cyc_when.cyt_when = pt->start + pt->period; + pt->kvm_cyc_when.cyt_interval = pt->period; + } else { + pt->kvm_cyc_when.cyt_when = pt->start + pt->period; + pt->kvm_cyc_when.cyt_when = CY_INFINITY; + } + + pt->kvm_cyclic_id = cyclic_add(&pt->kvm_cyc_handler, &pt->kvm_cyc_when); + pt->intervals = 0; + pt->active = 1; + mutex_exit(&cpu_lock); +} + +static void +pit_load_count(struct kvm *kvm, int channel, uint32_t val) +{ + struct kvm_kpit_state *ps = &kvm->arch.vpit->pit_state; + + ASSERT(mutex_owned(&ps->lock)); + + /* + * The largest possible initial count is 0; this is equivalent + * to 216 for binary counting and 104 for BCD counting. + */ + if (val == 0) + val = 0x10000; + + ps->channels[channel].count = val; + + if (channel != 0) { + ps->channels[channel].count_load_time = gethrtime(); + return; + } + + /* + * Two types of timer + * mode 1 is one shot, mode 2 is period, otherwise del timer + */ + switch (ps->channels[0].mode) { + case 0: + case 1: + /* FIXME: enhance mode 4 precision */ + case 4: + if (!(ps->flags & KVM_PIT_FLAGS_HPET_LEGACY)) + create_pit_timer(ps, val, 0); + break; + case 2: + case 3: + if (!(ps->flags & KVM_PIT_FLAGS_HPET_LEGACY)) + create_pit_timer(ps, val, 1); + break; + default: + destroy_pit_timer(&ps->pit_timer); + } +} + +static struct kvm_pit * +dev_to_pit(struct kvm_io_device *dev) +{ + return ((struct kvm_pit *)(((caddr_t)dev) - + offsetof(struct kvm_pit, dev))); +} + +static struct kvm_pit * +speaker_to_pit(struct kvm_io_device *dev) +{ + struct kvm_pit *pit = (struct kvm_pit *)(((caddr_t)dev) - + offsetof(struct kvm_pit, speaker_dev)); + + return (pit); +} + +static int +pit_in_range(gpa_t addr) +{ + return ((addr >= KVM_PIT_BASE_ADDRESS) && + (addr < KVM_PIT_BASE_ADDRESS + KVM_PIT_MEM_LENGTH)); +} + +static int +pit_ioport_write(struct kvm_io_device *this, + gpa_t addr, int len, const void *data) +{ + struct kvm_pit *pit = dev_to_pit(this); + struct kvm_kpit_state *pit_state = &pit->pit_state; + struct kvm *kvm = pit->kvm; + int channel, access; + struct kvm_kpit_channel_state *s; + uint32_t val = *(uint32_t *) data; + + if (!pit_in_range(addr)) + return (-EOPNOTSUPP); + + val &= 0xff; + addr &= KVM_PIT_CHANNEL_MASK; + + mutex_enter(&pit_state->lock); + + if (addr == 3) { + channel = val >> 6; + if (channel == 3) { + /* Read-Back Command. */ + for (channel = 0; channel < 3; channel++) { + s = &pit_state->channels[channel]; + if (val & (2 << channel)) { + if (!(val & 0x20)) + pit_latch_count(kvm, channel); + if (!(val & 0x10)) + pit_latch_status(kvm, channel); + } + } + } else { + /* Select Counter . */ + s = &pit_state->channels[channel]; + access = (val >> 4) & KVM_PIT_CHANNEL_MASK; + if (access == 0) { + pit_latch_count(kvm, channel); + } else { + s->rw_mode = access; + s->read_state = access; + s->write_state = access; + s->mode = (val >> 1) & 7; + if (s->mode > 5) + s->mode -= 4; + s->bcd = val & 1; + } + } + } else { + /* Write Count. */ + s = &pit_state->channels[addr]; + switch (s->write_state) { + default: + case RW_STATE_LSB: + pit_load_count(kvm, addr, val); + break; + case RW_STATE_MSB: + pit_load_count(kvm, addr, val << 8); + break; + case RW_STATE_WORD0: + s->write_latch = val; + s->write_state = RW_STATE_WORD1; + break; + case RW_STATE_WORD1: + pit_load_count(kvm, addr, s->write_latch | (val << 8)); + s->write_state = RW_STATE_WORD0; + break; + } + } + + mutex_exit(&pit_state->lock); + return (0); +} + +static int +pit_ioport_read(struct kvm_io_device *this, gpa_t addr, int len, void *data) +{ + struct kvm_pit *pit = dev_to_pit(this); + struct kvm_kpit_state *pit_state = &pit->pit_state; + struct kvm *kvm = pit->kvm; + int ret, count; + struct kvm_kpit_channel_state *s; + + if (!pit_in_range(addr)) + return (-EOPNOTSUPP); + + addr &= KVM_PIT_CHANNEL_MASK; + if (addr == 3) + return (0); + + s = &pit_state->channels[addr]; + + mutex_enter(&pit_state->lock); + + if (s->status_latched) { + s->status_latched = 0; + ret = s->status; + } else if (s->count_latched) { + switch (s->count_latched) { + default: + case RW_STATE_LSB: + ret = s->latched_count & 0xff; + s->count_latched = 0; + break; + case RW_STATE_MSB: + ret = s->latched_count >> 8; + s->count_latched = 0; + break; + case RW_STATE_WORD0: + ret = s->latched_count & 0xff; + s->count_latched = RW_STATE_MSB; + break; + } + } else { + switch (s->read_state) { + default: + case RW_STATE_LSB: + count = pit_get_count(kvm, addr); + ret = count & 0xff; + break; + case RW_STATE_MSB: + count = pit_get_count(kvm, addr); + ret = (count >> 8) & 0xff; + break; + case RW_STATE_WORD0: + count = pit_get_count(kvm, addr); + ret = count & 0xff; + s->read_state = RW_STATE_WORD1; + break; + case RW_STATE_WORD1: + count = pit_get_count(kvm, addr); + ret = (count >> 8) & 0xff; + s->read_state = RW_STATE_WORD0; + break; + } + } + + if (len > sizeof (ret)) + len = sizeof (ret); + + memcpy(data, (char *)&ret, len); + + mutex_exit(&pit_state->lock); + + return (0); +} + +static int +speaker_ioport_write(struct kvm_io_device *this, gpa_t addr, int len, + const void *data) +{ + struct kvm_pit *pit = speaker_to_pit(this); + struct kvm_kpit_state *pit_state = &pit->pit_state; + struct kvm *kvm = pit->kvm; + uint32_t val = *(uint32_t *) data; + + if (addr != KVM_SPEAKER_BASE_ADDRESS) + return (-EOPNOTSUPP); + + mutex_enter(&pit_state->lock); + pit_state->speaker_data_on = (val >> 1) & 1; + pit_set_gate(kvm, 2, val & 1); + mutex_exit(&pit_state->lock); + + return (0); +} + +static int +speaker_ioport_read(struct kvm_io_device *this, gpa_t addr, int len, void *data) +{ + struct kvm_pit *pit = speaker_to_pit(this); + struct kvm_kpit_state *pit_state = &pit->pit_state; + struct kvm *kvm = pit->kvm; + unsigned int refresh_clock; + int ret; + + if (addr != KVM_SPEAKER_BASE_ADDRESS) + return (-EOPNOTSUPP); + + /* Refresh clock toggles at about 15us. We approximate as 2^14ns. */ +#ifdef XXX + refresh_clock = ((unsigned int)ktime_to_ns(ktime_get()) >> 14) & 1; +#else + refresh_clock = ((unsigned int)gethrtime() >> 14) & 1; + XXX_KVM_PROBE; +#endif + + mutex_enter(&pit_state->lock); + ret = ((pit_state->speaker_data_on << 1) | pit_get_gate(kvm, 2) | + (pit_get_out(kvm, 2) << 5) | (refresh_clock << 4)); + + if (len > sizeof (ret)) + len = sizeof (ret); + + memcpy(data, (char *)&ret, len); + mutex_exit(&pit_state->lock); + return (0); +} + +void +kvm_pit_reset(struct kvm_pit *pit) +{ + int i; + struct kvm_kpit_channel_state *c; + + mutex_enter(&pit->pit_state.lock); + pit->pit_state.flags = 0; + for (i = 0; i < 3; i++) { + c = &pit->pit_state.channels[i]; + c->mode = 0xff; + c->gate = (i != 2); + pit_load_count(pit->kvm, i, 0); + } + mutex_exit(&pit->pit_state.lock); + + pit->pit_state.pit_timer.pending = 0; /* XXX need protection? */ + pit->pit_state.irq_ack = 1; +} + +static void +pit_mask_notifer(struct kvm_irq_mask_notifier *kimn, int mask) +{ + struct kvm_pit *pit = (struct kvm_pit *)(((caddr_t)kimn) - + offsetof(struct kvm_pit, mask_notifier)); + + if (!mask) { +#ifdef XXX + atomic_set(&pit->pit_state.pit_timer.pending, 0); +#else + pit->pit_state.pit_timer.pending = 0; + XXX_KVM_PROBE; +#endif + pit->pit_state.irq_ack = 1; + } +} + +static const struct kvm_io_device_ops pit_dev_ops = { + .read = pit_ioport_read, + .write = pit_ioport_write, +}; + +static const struct kvm_io_device_ops speaker_dev_ops = { + .read = speaker_ioport_read, + .write = speaker_ioport_write, +}; + +/* Caller must hold slots_lock */ +struct kvm_pit * +kvm_create_pit(struct kvm *kvm, uint32_t flags) +{ + struct kvm_pit *pit; + struct kvm_kpit_state *pit_state; + int ret; + + pit = kmem_zalloc(sizeof (struct kvm_pit), KM_SLEEP); + + pit->irq_source_id = kvm_request_irq_source_id(kvm); + + if (pit->irq_source_id < 0) { + kmem_free(pit, sizeof (struct kvm_pit)); + return (NULL); + } + + mutex_init(&pit->pit_state.lock, NULL, MUTEX_DRIVER, 0); + mutex_enter(&pit->pit_state.lock); +#ifdef XXX + raw_spin_lock_init(&pit->pit_state.inject_lock); +#else + XXX_KVM_SYNC_PROBE; + mutex_init(&pit->pit_state.inject_lock, NULL, MUTEX_DRIVER, 0); +#endif + kvm->arch.vpit = pit; + pit->kvm = kvm; + + pit_state = &pit->pit_state; + pit_state->pit = pit; +#ifdef XXX + hrtimer_init(&pit_state->pit_timer.timer, + CLOCK_MONOTONIC, HRTIMER_MODE_ABS); +#else + XXX_KVM_PROBE; +#endif + + pit_state->irq_ack_notifier.gsi = 0; + pit_state->irq_ack_notifier.irq_acked = kvm_pit_ack_irq; + + kvm_register_irq_ack_notifier(kvm, &pit_state->irq_ack_notifier); + + pit_state->pit_timer.reinject = 1; + pit_state->pit_timer.active = 0; + + mutex_exit(&pit->pit_state.lock); + + kvm_pit_reset(pit); + + pit->mask_notifier.func = pit_mask_notifer; + kvm_register_irq_mask_notifier(kvm, 0, &pit->mask_notifier); + + kvm_iodevice_init(&pit->dev, &pit_dev_ops); + ret = kvm_io_bus_register_dev(kvm, KVM_PIO_BUS, &pit->dev); + if (ret < 0) + goto fail; + + if (flags & KVM_PIT_SPEAKER_DUMMY) { + kvm_iodevice_init(&pit->speaker_dev, &speaker_dev_ops); + ret = kvm_io_bus_register_dev(kvm, + KVM_PIO_BUS, &pit->speaker_dev); + + if (ret < 0) + goto fail_unregister; + } + + return (pit); + +fail_unregister: +#ifdef XXX + kvm_io_bus_unregister_dev(kvm, KVM_PIO_BUS, &pit->dev); +#else + XXX_KVM_PROBE; +#endif +fail: +#ifdef XXX + kvm_unregister_irq_mask_notifier(kvm, 0, &pit->mask_notifier); + kvm_unregister_irq_ack_notifier(kvm, &pit_state->irq_ack_notifier); + kvm_free_irq_source_id(kvm, pit->irq_source_id); + kmem_free(pit, sizeof (struct kvm_pit)); +#else + XXX_KVM_PROBE; +#endif + return (NULL); +} + +void +kvm_free_pit(struct kvm *kvmp) +{ + struct kvm_timer *kptp; + + if (kvmp->arch.vpit == NULL) + return; + + mutex_enter(&kvmp->arch.vpit->pit_state.lock); + kvm_unregister_irq_mask_notifier(kvmp, 0, + &kvmp->arch.vpit->mask_notifier); + kvm_unregister_irq_ack_notifier(kvmp, + &kvmp->arch.vpit->pit_state.irq_ack_notifier); + mutex_exit(&kvmp->arch.vpit->pit_state.lock); + + mutex_enter(&cpu_lock); + kptp = &kvmp->arch.vpit->pit_state.pit_timer; + if (kptp->active) + cyclic_remove(kptp->kvm_cyclic_id); + mutex_exit(&cpu_lock); + mutex_destroy(&kvmp->arch.vpit->pit_state.lock); + kvm_free_irq_source_id(kvmp, kvmp->arch.vpit->irq_source_id); + kmem_free(kvmp->arch.vpit, sizeof (struct kvm_pit)); + kvmp->arch.vpit = NULL; +} + static void __inject_pit_timer_intr(struct kvm *kvm) { @@ -119,29 +813,3 @@ kvm_inject_pit_timer_irqs(struct kvm_vcpu *vcpu) __inject_pit_timer_intr(kvm); } } - -void -kvm_free_pit(struct kvm *kvmp) -{ - struct kvm_timer *kptp; - - if (kvmp->arch.vpit == NULL) - return; - - mutex_enter(&kvmp->arch.vpit->pit_state.lock); - kvm_unregister_irq_mask_notifier(kvmp, 0, - &kvmp->arch.vpit->mask_notifier); - kvm_unregister_irq_ack_notifier(kvmp, - &kvmp->arch.vpit->pit_state.irq_ack_notifier); - mutex_exit(&kvmp->arch.vpit->pit_state.lock); - - mutex_enter(&cpu_lock); - kptp = &kvmp->arch.vpit->pit_state.pit_timer; - if (kptp->active) - cyclic_remove(kptp->kvm_cyclic_id); - mutex_exit(&cpu_lock); - mutex_destroy(&kvmp->arch.vpit->pit_state.lock); - kvm_free_irq_source_id(kvmp, kvmp->arch.vpit->irq_source_id); - kmem_free(kvmp->arch.vpit, sizeof (struct kvm_pit)); - kvmp->arch.vpit = NULL; -} -- cgit v1.2.3