[Top][All Lists]

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Qemu-ppc] Proposal PCI/PCIe device placement on PAPR guests

From: Michael Roth
Subject: Re: [Qemu-ppc] Proposal PCI/PCIe device placement on PAPR guests
Date: Thu, 12 Jan 2017 10:35:37 -0600
User-agent: alot/0.3.6

Quoting Laine Stump (2017-01-12 08:52:10)
> On 01/12/2017 05:31 AM, Andrea Bolognani wrote:
> > On Mon, 2017-01-09 at 10:46 +1100, David Gibson wrote:
> >>>>     * To allow for hotplugged devices, libvirt should also add a number
> >>>>       of additional, empty vPHBs (the PAPR spec allows for hotplug of
> >>>>       PHBs, but this is not yet implemented in qemu).
> >>>
> >>> "A number" here will have to mean "one", same number of
> >>> empty PCIe Root Ports libvirt will add to a newly-defined
> >>> q35 guest.
> >>
> >> Umm.. why?
> >
> > Because some applications using libvirt would inevitably
> > start relying on the fact that such spare PHBs are
> > available, locking us into providing at least the same
> > number forever. In other words, increasing the amount at
> > a later time is always possible, but decreasing it isn't.
> > We did the same when we started automatically adding PCIe
> > Root Ports to q35 machines.
> >
> > The rationale is that having a single spare hotpluggable
> > slot is extremely convenient for basic usage, eg. a simple
> > guest created by someone who's not necessarily very
> > familiar with virtualization; on the other hand, if you
> > are actually deploying in production you ought to conduct
> > proper capacity planning and figure out in advance how
> > many devices you're likely to need to hotplug throughout
> > the guest's life.
> And of course the reason we don't want to add "too many" extra 
> controllers by default is so that we don't end up with *all* guests 
> burdened with extra hardware they don't need or want. The libguestfs 
> appliance is one example of a libvirt consumer that definitely doesn't 
> want extra baggage in its guests - guest startup time is very important 
> to libguestfs, so any addition to the hardware list is looked upon with 
> disappointment.
> >
> > Of course this all will be moot once we can hotplug PHBs :)
> Will the guest OSes handle that properly? I remember being told that 

I believe on pseries we *do* scan for devices on the PHB as part of
bringing the PHB online in the hotplug path. But I'm not sure that
matters (see below).

> Linux, for example, doesn't scan the new bus for devices when a new 
> controller is added, making it pointless to hotplug a PCI controller (as 
> usual, it could be that I'm remembering incorrectly...)

Wouldn't that only be an issue if we hotplugged a PHB that already had
PCI devices on the bus? That only seems possible if we had a way to
signal phb hotplug *after* we've hotplugged some PCI devices on the bus,
which means we'd need some interface to trigger hotplug  beyond the
standard 'device_add' calls, e.g.:

  device_add spapr-pci-host-bridge,hotplug-deferred=true,id=phb2,index=2
  device_add virtio-net-pci,bus=phb2.0,...,hotplug-deferred=true
  device_signal_hotplug phb2

That's actually akin to how it's normally done on pHyp (not only for PHB
hotplug, but for PCI hotplug in general, which is why this could be
reasonably expected to work on pseries guests), but it seems quite a bit
different from how we'd normally handle this on kvm, which I think would
be something more like:

  device_add spapr-pci-host-bridge,id=phb2,index=2
  <wait for hotplug completion event>
  device_add virtio-net-pci,bus=phb2.0,...

In which case it doesn't really matter if the guest scans the bus at
hotplug time or not. Is there some other scenario where this might

reply via email to

[Prev in Thread] Current Thread [Next in Thread]