Re: [PATCH v3 02/15] dt/bindings: Update binding for PM domain idle states

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 






On 19/08/16 19:10, Kevin Hilman wrote:
Sudeep Holla <sudeep.holla@xxxxxxx> writes:

[...]

In general whatever binding we come up must not just address OS
coordinated mode. Also I was thinking to have better coverage in the
description by having a bit more complex system like:

cluster0
	CLUSTER_RET(Retention)
	CLUSTER_PG(Power Gate)
	core0
		CORE_RET
		CORE_PG
	core1
		CORE_RET
		CORE_PG

Also, remember that a power domain may contain more than just CPUs, so
this will also need to handle things like:

	device0..N
        	DEV_CLK_GATE
                DEV_RET
                DEV_PG

So, as (I think) Lina was trying to say, including CPU idle states
inside domain idles states doesn't really scale well because it would
also imply domain states would also include device idle states.

IMO, the device-specific states belong in the device nodes, and that
includes CPUs.


OK, IIUC we don't have device idle states binding today, so we are not
breaking anything there. Can you elaborate on the issue you see if we
just have domain idle-states ? Is it because we currently create genpd
domain for each entry ?

If a CPU/Device can be enter idle-state(s) it means that it is in a
power domain on its own, so I don't see any issue in such representation.

It's up to the domain (genpd) governor to look at *all* devices in the
domain, check their state and make a domain-wide decision.


Lets not mix the current genpd implementation in the kernel into this
discussion for simplicity. How is the implementation in the kernel today
and what can be done is a separate topic.

What this discussion should aim at is to present the idle states in the
system in the device tree so that it address the issues we have
currently and extensible in near future with any compatibility issues.

The tricky part remains, IMO, the mapping between device/CPU states and
allowable domain states.

As was suggested earlier, a good potential starting point would be that
all devices/CPUs would need to be in their deepest state before the
domain would make any decisions.  While that leaves soem power savings
on the table, it maps well to how genpd works today with only on/off
states and could be extended with more complicated governors down the
road.


Agreed.

Some example below for discussion, feel free to add more cases.

--
Regards,
Sudeep


--->8

1. Dual cluster with 2 CPUs in each cluster with powerdown at both CPU and
   cluster level

  idle-states {
    CPU_SLEEP_0: cpu-sleep-0 {
      ...
      entry-latency-us = <300>;
      ...
    };
    CLUSTER_SLEEP_0: cluster-sleep-0 {
      ...
      entry-latency-us = <300>;
      ...
    };
  };

  cpu@0 {
    ...
    /*
     * implentation may ignore cpu-idle-states if power-domains
     * has idle-states, DT's may have both for backward compatibility
     */
    cpu-idle-states = <&CPU_SLEEP_0 &CLUSTER_SLEEP_0>;
    power-domains = <&CPU_0_1_PD>;
    ...
  };

  cpu@1 {
    ...
    cpu-idle-states = <&CPU_SLEEP_0 &CLUSTER_SLEEP_0>;
    power-domains = <&CPU_0_1_PD>;
    ...
  };

  cpu@100 {
    ...
    cpu-idle-states = <&CPU_SLEEP_0 &CLUSTER_SLEEP_0>;
    power-domains = <&CPU_1_0_PD>;
    ...
  };

  cpu@101 {
    ...
    cpu-idle-states = <&CPU_SLEEP_0 &CLUSTER_SLEEP_0>;
    power-domains = <&CPU_1_1_PD>;
    ...
  };

  power-domains {
    CLUSTER_0_PD: cluster-0-pd {
      #power-domain-cells = <0>;
      domain-idle-states = <&CLUSTER_SLEEP_0>;
    };
    CPU_0_0_PD: cpu-0-0-pd@0 {
      #power-domain-cells = <0>;
      domain-idle-states = <&CPU_SLEEP_0>;
      power-domains = <&CLUSTER_0_PD>;
    };
    CPU_0_1_PD: cpu-0-1-pd@1 {
      #power-domain-cells = <0>;
      domain-idle-states = <&CPU_SLEEP_0>;
      power-domains = <&CLUSTER_0_PD>;
    };
    CLUSTER_1_PD: cluster-1-pd {
      #power-domain-cells = <0>;
      domain-idle-states = <&CLUSTER_SLEEP_0>;
    };
    CPU_1_0_PD: cpu-1-0-pd@0 {
      #power-domain-cells = <0>;
      domain-idle-states = <&CPU_SLEEP_0>;
      power-domains = <&CLUSTER_1_PD>;
    };
    CPU_1_1_PD: cpu-1-1-pd@1 {
      #power-domain-cells = <0>;
      domain-idle-states = <&CPU_SLEEP_0>;
      power-domains = <&CLUSTER_1_PD>;
    };
  };

2. Dual cluster with 2 CPUs in each cluster with retention and powerdown at both
   CPU and cluster level

  idle-states {
    CPU_SLEEP_0: cpu-sleep-0 { /* Retention */
      ...
      entry-latency-us = <100>;
      ...
    };
    CPU_SLEEP_1: cpu-sleep-1 { /* Power-down */
      ...
      entry-latency-us = <500>;
      ...
    };
    CLUSTER_SLEEP_0: cluster-sleep-0 { /* Retention */
      ...
      entry-latency-us = <300>;
      ...
    };
    CLUSTER_SLEEP_1: cluster-sleep-1 {/* Power-down */
      ...
      entry-latency-us = <1000>;
      ...
    };
  };

  cpu@0 {
    ...
    power-domains = <&CPU_0_1_PD>;
    ...
  };

  cpu@1 {
    ...
    power-domains = <&CPU_0_1_PD>;
    ...
  };

  cpu@100 {
    ...
    power-domains = <&CPU_1_0_PD>;
    ...
  };

  cpu@101 {
    ...
    power-domains = <&CPU_1_1_PD>;
    ...
  };

  power-domains {
    /*
     * Each cluster/core PD may point to different idle states,
     * it's all same here in the example to keep it short and
     * simple
     */
    CLUSTER_0_PD: cluster-0-pd {
      #power-domain-cells = <0>;
      domain-idle-states = <&CLUSTER_SLEEP_0 &CLUSTER_SLEEP_1>;
    };
    CPU_0_0_PD: cpu-0-0-pd@0 {
      #power-domain-cells = <0>;
      domain-idle-states = <&CPU_SLEEP_0 &CPU_SLEEP_1>;
      power-domains = <&CLUSTER_0_PD>;
    };
    CPU_0_1_PD: cpu-0-1-pd@1 {
      #power-domain-cells = <0>;
      domain-idle-states = <&CPU_SLEEP_0 &CPU_SLEEP_1>;
      power-domains = <&CLUSTER_0_PD>;
    };
    CLUSTER_1_PD: cluster-1-pd {
      #power-domain-cells = <0>;
      domain-idle-states = <&CLUSTER_SLEEP_0 &CLUSTER_SLEEP_1>;
    };
    CPU_1_0_PD: cpu-1-0-pd@0 {
      #power-domain-cells = <0>;
      domain-idle-states = <&CPU_SLEEP_0 &CPU_SLEEP_1>;
      power-domains = <&CLUSTER_1_PD>;
    };
    CPU_1_1_PD: cpu-1-1-pd@1 {
      #power-domain-cells = <0>;
      domain-idle-states = <&CPU_SLEEP_0 &CPU_SLEEP_1>;
      power-domains = <&CLUSTER_1_PD>;
    };
  };

3. Dual cluster with 2 CPUs in each cluster with retention and powerdown at
   just cluster level

  idle-states {
    CLUSTER_SLEEP_0: cluster-sleep-0 { /* Retention */
      ...
      entry-latency-us = <300>;
      ...
    };
    CLUSTER_SLEEP_1: cluster-sleep-1 {/* Power-down */
      ...
      entry-latency-us = <1000>;
      ...
    };
  };

  cpu@0 {
    ...
    power-domains = <&CLUSTER_0_PD>;
    ...
  };

  cpu@1 {
    ...
    power-domains = <&CLUSTER_0_PD>;
    ...
  };

  cpu@100 {
    ...
    power-domains = <&CLUSTER_1_PD>;
    ...
  };

  cpu@101 {
    ...
    power-domains = <&CLUSTER_1_PD>;
    ...
  };

  power-domains {
    CLUSTER_0_PD: cluster-0-pd {
      #power-domain-cells = <0>;
      domain-idle-states = <&CLUSTER_SLEEP_0 &CLUSTER_SLEEP_1>;
    };
    CLUSTER_1_PD: cluster-1-pd {
      #power-domain-cells = <0>;
      domain-idle-states = <&CLUSTER_SLEEP_0 &CLUSTER_SLEEP_1>;
    };
  };

4. 4 devices sharing the power domain.

  idle-states {
    /*
     * Device idle states may differ from CPU idle states in terms
     * of the list of properties
     */
    DEVPD_SLEEP_0: devpd-sleep-0 { /* Retention */
      ...
      entry-latency-us = <300>;
      ...
    };
    DEVPD_SLEEP_1: devpd-sleep-1 {/* Power-down */
      ...
      entry-latency-us = <1000>;
      ...
    };
  };

  dev@0 {
    ...
    power-domains = <&DEV_PD_0>;
    ...
  };

  dev@1 {
    ...
    power-domains = <&DEV_PD_0>;
    ...
  };

  dev@2 {
    ...
    power-domains = <&DEV_PD_0>;
    ...
  };

  dev@3 {
    ...
    power-domains = <&DEV_PD_0>;
    ...
  };

  power-domains {
    DEV_PD_0: device-pd-0 {
      #power-domain-cells = <0>;
      domain-idle-states = <&DEVPD_SLEEP_0 &DEVPD_SLEEP_1>;
    };
  };

5. 4 devices sharing the power domain + another device sharing the power
   domain but has it's own sub-domain

  idle-states {
    DEVPD_0_SLEEP_0: devpd-sleep-0 { /* Retention */
      ...
      entry-latency-us = <300>;
      ...
    };
    DEVPD_0_SLEEP_1: devpd-sleep-1 {/* Power-down */
      ...
      entry-latency-us = <1000>;
      ...
    };
    DEVPD_1_SLEEP_0: devpd-sleep-0 { /* Retention */
      ...
      entry-latency-us = <300>;
      ...
    };
    DEVPD_1_SLEEP_1: devpd-sleep-1 {/* Power-down */
      ...
      entry-latency-us = <1000>;
      ...
    };
  };

  dev@0 {
    ...
    power-domains = <&DEV_PD_0>;
    ...
  };

  dev@1 {
    ...
    power-domains = <&DEV_PD_0>;
    ...
  };

  dev@2 {
    ...
    power-domains = <&DEV_PD_0>;
    ...
  };

  dev@3 {
    ...
    power-domains = <&DEV_PD_0>;
    ...
  };

  dev@4 {
    ...
    power-domains = <&DEV_PD_1>;
    ...
  };

  power-domains {
    DEV_PD_0: device-pd-0 {
      #power-domain-cells = <0>;
      domain-idle-states = <&DEVPD_0_SLEEP_0 &DEVPD_0_SLEEP_1>;
    };
    DEV_PD_1: device-pd-1 {
      #power-domain-cells = <0>;
      power-domains = <&DEV_PD_0>;
      domain-idle-states = <&DEVPD_1_SLEEP_0 &DEVPD_1_SLEEP_1>;
    };
  };
--
To unsubscribe from this list: send the line "unsubscribe devicetree" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [Device Tree Compilter]     [Device Tree Spec]     [Linux Driver Backports]     [Video for Linux]     [Linux USB Devel]     [Linux PCI Devel]     [Linux Audio Users]     [Linux Kernel]     [Linux SCSI]     [XFree86]     [Yosemite Backpacking]
  Powered by Linux