On Tuesday 11 November 2014 15:01:28 Sudeep Holla wrote: > On 11/11/14 13:18, Ashwin Chaugule wrote: > > On 11 November 2014 05:30, Sudeep Holla <sudeep.holla@xxxxxxx> wrote: > >> On 07/11/14 14:52, Ashwin Chaugule wrote: > >>> +static struct mbox_chan pcc_mbox_chan[MAX_PCC_SUBSPACES]; > >> > >> > >> Really, do you want to allocate 256 structures of mbox_chan even on > >> systems with just 1 - 2 channels(typically that would be the case) ? > >> > > > > Spec says, max of 256. Easier to support it this way. > > > > I tend to disagree, you can allocate dynamically. And use exactly same > logic you use in get_subspace_id to populate con_priv later. When you > parse PCC Subspace, you can just validate header and not record them in > pcc_mbox_chan. You can get the exact count in acpi_table_parse_entries > and then do the required allocation. Right, I think that would be best. > >>> + > >>> + count = acpi_table_parse_entries(ACPI_SIG_PCCT, > >>> + sizeof(struct acpi_table_pcct), > >> > >> > >> s/struct acpi_table_pcct/*pcct_tbl/ > >> > >> In general, the interrupt part of the PCCT SS is not considered in this > >> patch. It's better to see/discuss how that can be fit into the mailbox > >> framework, though it could be follow up patch. > > > > The IRQ part of the spec seems to be under discussion (single irq per > > subspace / common IRQ across all) and as you may be aware we're > > working on trying it out on Juno. That'll guide the design. What I > > have here is good enough to start off with and has been tested. I dont > > think we should have a problem using the mailbox API for asyn tx > > though, but I'd really^n prefer if we get something out there first. > > > > That's the different and still under discussion. But you need to support > Type 1 subspace as it stands in ACPI v5.1 Why? We should only implement whatever is required to support existing hardware, not because something is in the spec. Arnd -- To unsubscribe from this list: send the line "unsubscribe linux-acpi" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html