Re: [PATCH] Fix bugs in runq

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



于 2012年08月30日 02:30, Dave Anderson 写道:
> 
> 
> ----- Original Message -----
> 
>>> Another question re: your patch -- is it possible to have a "depth" greater
>>> than 1?
>>
>> Yes, "depth" could be greater than 1, see the example below:
>>
>> CPU 0 RUNQUEUE: ffff880028216680
>>   CURRENT: PID: 17085  TASK: ffff880137c63540  COMMAND: "bash"
>>   RT PRIO_ARRAY: ffff880028216808                                        <-- depth = 0
>>      [  0] PID: 17129  TASK: ffff880037aeaaa0  COMMAND: "rtloop99"
>>            PID: 2832   TASK: ffff88013b09cae0  COMMAND: "rtkit-daemon"
>>            PID: 6      TASK: ffff88013d7c6080  COMMAND: "watchdog/0"
>>      [  1] GROUP RT PRIO_ARRAY: ffff88002ca65000                         <-- depth = 1
>>               [  1] GROUP RT PRIO_ARRAY: ffff880015821000                <-- depth = 2
>>                        [  1] PID: 17126  TASK: ffff880135d2a040  COMMAND: "rtloop98"
>>               [ 98] PID: 17119  TASK: ffff88010190d500  COMMAND: "rtloop1"
>>                     PID: 17121  TASK: ffff88013bd27500  COMMAND: "rtloop1"
>>                     PID: 17120  TASK: ffff88010190caa0  COMMAND: "rtloop1"
>>   CFS RB_ROOT: ffff880028216718
> ...
> 
>> Hmm, I think the depth could not be that big. So how do you think this
>> kind of output.
>>
>> The attached patch just changed "CHILD" to "GROUP".
> 
> Interesting -- how did you set up the depth-greater-than-one scenario?

Below is my script to run a number of rt tasks in different cgroups:

#!/bin/bash

mkdir /cgroup/cpu/test1
echo 850000 > /cgroup/cpu/test1/cpu.rt_runtime_us

./rtloop1 &
echo $! > /cgroup/cpu/test1/tasks
./rtloop1 &
echo $! > /cgroup/cpu/test1/tasks
./rtloop1 &
echo $! > /cgroup/cpu/test1/tasks
./rtloop98 &
echo $! > /cgroup/cpu/test1/tasks
./rtloop45 &
echo $! > /cgroup/cpu/test1/tasks
./rtloop99 &
echo $! > /cgroup/cpu/test1/tasks

mkdir /cgroup/cpu/test1/test11
echo 550000 > /cgroup/cpu/test1/test11/cpu.rt_runtime_us

./rtloop98 &
echo $! > /cgroup/cpu/test1/test11/tasks
./rtloop99 &
echo $! > /cgroup/cpu/test1/test11/tasks

./rtloop98 &
./rtloop99 &

> 
> Anyway, given that it is possible, let's at least tighten up the output display
> by changing each "9 * depth" usage to be "6 * depth".  That should alter
> your example output to look like this:
> 
> CPU 0 RUNQUEUE: ffff880028216680
>   CURRENT: PID: 17085  TASK: ffff880137c63540  COMMAND: "bash"
>   RT PRIO_ARRAY: ffff880028216808                                       
>      [  0] PID: 17129  TASK: ffff880037aeaaa0  COMMAND: "rtloop99"
>            PID: 2832   TASK: ffff88013b09cae0  COMMAND: "rtkit-daemon"
>            PID: 6      TASK: ffff88013d7c6080  COMMAND: "watchdog/0"
>      [  1] GROUP RT PRIO_ARRAY: ffff88002ca65000                       
>            [  1] GROUP RT PRIO_ARRAY: ffff880015821000                
>                  [  1] PID: 17126  TASK: ffff880135d2a040  COMMAND: "rtloop98"
>            [ 98] PID: 17119  TASK: ffff88010190d500  COMMAND: "rtloop1"
>                  PID: 17121  TASK: ffff88013bd27500  COMMAND: "rtloop1"
>                  PID: 17120  TASK: ffff88010190caa0  COMMAND: "rtloop1"
>   CFS RB_ROOT: ffff880028216718
>      ...

Hmm, This kind of output looks not that understandable easily...
for example:
  RT PRIO_ARRAY: ffff880028296808
     [  0] GROUP RT PRIO_ARRAY: ffff880103ded800
           [  0] GROUP RT PRIO_ARRAY: ffff88011ae70800
                 [  0] PID: 17127  TASK: ffff8800378f6040  COMMAND: "rtloop99"
                 PID: 17124  TASK: ffff8800a9592ae0  COMMAND: "rtloop99"
           [  1] PID: 17122  TASK: ffff88011aec3500  COMMAND: "rtloop98"
           [ 54] PID: 17123  TASK: ffff88013b414ae0  COMMAND: "rtloop45"
           PID: 10     TASK: ffff88013cc2cae0  COMMAND: "watchdog/1"
           PID: 7      TASK: ffff88013d7ef500  COMMAND: "migration/1"
     [  1] PID: 17128  TASK: ffff880139761540  COMMAND: "rtloop98"

This output looks confusing, it seems that task 10 and 7 are in group
ffff880103ded800

I think we can avoid this by two ways:
1. 6 * depth --> 8 * depth, proper indentation.
     RT PRIO_ARRAY: ffff880028296808
     [  0] GROUP RT PRIO_ARRAY: ffff880103ded800
             [  0] GROUP RT PRIO_ARRAY: ffff88011ae70800
                     [  0] PID: 17127  TASK: ffff8800378f6040  COMMAND: "rtloop99"
                   PID: 17124  TASK: ffff8800a9592ae0  COMMAND: "rtloop99"
             [  1] PID: 17122  TASK: ffff88011aec3500  COMMAND: "rtloop98"
             [ 54] PID: 17123  TASK: ffff88013b414ae0  COMMAND: "rtloop45"
           PID: 10     TASK: ffff88013cc2cae0  COMMAND: "watchdog/1"
           PID: 7      TASK: ffff88013d7ef500  COMMAND: "migration/1"
     [  1] PID: 17128  TASK: ffff880139761540  COMMAND: "rtloop98"

2. print priority for all tasks,
     RT PRIO_ARRAY: ffff880028296808
     [  0] GROUP RT PRIO_ARRAY: ffff880103ded800
           [  0] GROUP RT PRIO_ARRAY: ffff88011ae70800
                 [  0] PID: 17127  TASK: ffff8800378f6040  COMMAND: "rtloop99"
           [  0] PID: 17124  TASK: ffff8800a9592ae0  COMMAND: "rtloop99"
           [  1] PID: 17122  TASK: ffff88011aec3500  COMMAND: "rtloop98"
           [ 54] PID: 17123  TASK: ffff88013b414ae0  COMMAND: "rtloop45"
     [  0] PID: 10     TASK: ffff88013cc2cae0  COMMAND: "watchdog/1"
     [  0] PID: 7      TASK: ffff88013d7ef500  COMMAND: "migration/1"
     [  1] PID: 17128  TASK: ffff880139761540  COMMAND: "rtloop98"

I prefer the second one, how do you think?
And I attached 2 patches, one is for way 1, and the other is for way 2.

> 
> And also, I'd prefer to not create the dangling "static int depth", 
> but rather to add a depth argument to dump_RT_prio_array(), where 
> dump_CFS_runqueues() passes a 0, and dump_RT_prio_array() passes
> "depth+1" to itself: 
> 
> static void
> dump_RT_prio_array(int depth, ulong k_prio_array, char *u_prio_array)
> {

......

> }
> 
> Can you verify that those changes work for you?

OK.

Thanks
Zhang Yanfei

>From c1330d466e459a680a967f374939d91794a940e7 Mon Sep 17 00:00:00 2001
From: zhangyanfei <zhangyanfei@xxxxxxxxxxxxxx>
Date: Thu, 30 Aug 2012 10:12:10 +0800
Subject: [PATCH] Fix rt not support group sched bug

Signed-off-by: zhangyanfei <zhangyanfei@xxxxxxxxxxxxxx>
---
 defs.h    |    2 ++
 symbols.c |    4 ++++
 task.c    |   52 ++++++++++++++++++++++++++++++++++++++++------------
 3 files changed, 46 insertions(+), 12 deletions(-)

diff --git a/defs.h b/defs.h
index 4a8e2e3..4af670d 100755
--- a/defs.h
+++ b/defs.h
@@ -1785,6 +1785,7 @@ struct offset_table {                    /* stash of commonly-used offsets */
 	long log_level;
 	long log_flags_level;
 	long timekeeper_xtime_sec;
+	long sched_rt_entity_my_q;
 };
 
 struct size_table {         /* stash of commonly-used sizes */
@@ -1919,6 +1920,7 @@ struct size_table {         /* stash of commonly-used sizes */
 	long msg_queue;
 	long log;
 	long log_level;
+	long rt_rq;
 };
 
 struct array_table {
diff --git a/symbols.c b/symbols.c
index 2646ff8..bbadd5e 100755
--- a/symbols.c
+++ b/symbols.c
@@ -8812,6 +8812,8 @@ dump_offset_table(char *spec, ulong makestruct)
 		OFFSET(log_level));
 	fprintf(fp, "               log_flags_level: %ld\n",
 		OFFSET(log_flags_level));
+	fprintf(fp, "          sched_rt_entity_my_q: %ld\n",
+		OFFSET(sched_rt_entity_my_q));
 
 	fprintf(fp, "\n                    size_table:\n");
 	fprintf(fp, "                          page: %ld\n", SIZE(page));
@@ -9027,6 +9029,8 @@ dump_offset_table(char *spec, ulong makestruct)
 		SIZE(log));
 	fprintf(fp, "                     log_level: %ld\n",
 		SIZE(log_level));
+	fprintf(fp, "                         rt_rq: %ld\n",
+		SIZE(rt_rq));
 
         fprintf(fp, "\n                   array_table:\n");
 	/*
diff --git a/task.c b/task.c
index 6e4cfec..824058c 100755
--- a/task.c
+++ b/task.c
@@ -67,7 +67,7 @@ static void dump_task_runq_entry(struct task_context *);
 static int dump_tasks_in_cfs_rq(ulong);
 static void dump_on_rq_tasks(void);
 static void dump_CFS_runqueues(void);
-static void dump_RT_prio_array(ulong, char *);
+static void dump_RT_prio_array(int, ulong, char *);
 static void task_struct_member(struct task_context *,unsigned int, struct reference *);
 static void signal_reference(struct task_context *, ulong, struct reference *);
 static void do_sig_thread_group(ulong);
@@ -7552,6 +7552,7 @@ dump_CFS_runqueues(void)
 
 	if (!VALID_STRUCT(cfs_rq)) {
 		STRUCT_SIZE_INIT(cfs_rq, "cfs_rq");
+		STRUCT_SIZE_INIT(rt_rq, "rt_rq");
 		MEMBER_OFFSET_INIT(rq_rt, "rq", "rt");
 		MEMBER_OFFSET_INIT(rq_nr_running, "rq", "nr_running");
 		MEMBER_OFFSET_INIT(task_struct_se, "task_struct", "se");
@@ -7562,6 +7563,8 @@ dump_CFS_runqueues(void)
 			"cfs_rq");
 		MEMBER_OFFSET_INIT(sched_entity_my_q, "sched_entity", 
 			"my_q");
+		MEMBER_OFFSET_INIT(sched_rt_entity_my_q, "sched_rt_entity",
+			"my_q");
 		MEMBER_OFFSET_INIT(sched_entity_on_rq, "sched_entity", "on_rq");
 		MEMBER_OFFSET_INIT(cfs_rq_rb_leftmost, "cfs_rq", "rb_leftmost");
 		MEMBER_OFFSET_INIT(cfs_rq_nr_running, "cfs_rq", "nr_running");
@@ -7629,7 +7632,7 @@ dump_CFS_runqueues(void)
 				OFFSET(cfs_rq_tasks_timeline));
 		}
 
-		dump_RT_prio_array(runq + OFFSET(rq_rt) + OFFSET(rt_rq_active), 
+		dump_RT_prio_array(0, runq + OFFSET(rq_rt) + OFFSET(rt_rq_active),
 			&runqbuf[OFFSET(rq_rt) + OFFSET(rt_rq_active)]);
 
 		fprintf(fp, "  CFS RB_ROOT: %lx\n", (ulong)root);
@@ -7649,7 +7652,7 @@ dump_CFS_runqueues(void)
 }
 
 static void
-dump_RT_prio_array(ulong k_prio_array, char *u_prio_array)
+dump_RT_prio_array(int depth, ulong k_prio_array, char *u_prio_array)
 {
 	int i, c, tot, cnt, qheads;
 	ulong offset, kvaddr, uvaddr;
@@ -7657,8 +7660,11 @@ dump_RT_prio_array(ulong k_prio_array, char *u_prio_array)
         struct list_data list_data, *ld;
 	struct task_context *tc;
 	ulong *tlist;
+	ulong my_q, task_addr;
+	char *rt_rq_buf;
 
-	fprintf(fp, "  RT PRIO_ARRAY: %lx\n",  k_prio_array);
+	if (!depth)
+		fprintf(fp, "  RT PRIO_ARRAY: %lx\n",  k_prio_array);
 
         qheads = (i = ARRAY_LENGTH(rt_prio_array_queue)) ?
                 i : get_array_length("rt_prio_array.queue", NULL, SIZE(list_head));
@@ -7678,14 +7684,11 @@ dump_RT_prio_array(ulong k_prio_array, char *u_prio_array)
 		if ((list_head[0] == kvaddr) && (list_head[1] == kvaddr))
 			continue;
 
-		fprintf(fp, "     [%3d] ", i);
-
 		BZERO(ld, sizeof(struct list_data));
 		ld->start = list_head[0];
 		if (VALID_MEMBER(task_struct_rt) &&
 		    VALID_MEMBER(sched_rt_entity_run_list))
-			ld->list_head_offset = OFFSET(task_struct_rt) + 
-				OFFSET(sched_rt_entity_run_list);
+			ld->list_head_offset = OFFSET(sched_rt_entity_run_list);
 		else
 			ld->list_head_offset = OFFSET(task_struct_run_list);
 		ld->end = kvaddr;
@@ -7695,10 +7698,35 @@ dump_RT_prio_array(ulong k_prio_array, char *u_prio_array)
 		tlist = (ulong *)GETBUF((cnt) * sizeof(ulong));
 		cnt = retrieve_list(tlist, cnt);
 		for (c = 0; c < cnt; c++) {
-			if (!(tc = task_to_context(tlist[c])))
+			task_addr = tlist[c];
+			if (VALID_MEMBER(sched_rt_entity_my_q)) {
+				readmem(tlist[c] + OFFSET(sched_rt_entity_my_q),
+					KVADDR, &my_q, sizeof(ulong), "my_q",
+					FAULT_ON_ERROR);
+				if (my_q) {
+					rt_rq_buf = GETBUF(SIZE(rt_rq));
+					readmem(my_q, KVADDR, rt_rq_buf,
+						SIZE(rt_rq), "rt_rq",
+						FAULT_ON_ERROR);
+
+					INDENT(5 + 6 * depth);
+					fprintf(fp, "[%3d] ", i);
+					fprintf(fp, "GROUP RT PRIO_ARRAY: %lx\n",
+						my_q + OFFSET(rt_rq_active));
+					tot++;
+					dump_RT_prio_array(depth + 1,
+						my_q + OFFSET(rt_rq_active),
+						&rt_rq_buf[OFFSET(rt_rq_active)]);
+					continue;
+				} else {
+					task_addr -= OFFSET(task_struct_rt);
+				}
+			}
+			if (!(tc = task_to_context(task_addr)))
 				continue;
-			if (c)
-				INDENT(11);
+
+			INDENT(5 + 6 * depth);
+			fprintf(fp, "[%3d] ", i);
 			fprintf(fp, "PID: %-5ld  TASK: %lx  COMMAND: \"%s\"\n",
 				tc->pid, tc->task, tc->comm);
 			tot++;
@@ -7707,7 +7735,7 @@ dump_RT_prio_array(ulong k_prio_array, char *u_prio_array)
 	}
 
 	if (!tot) {
-		INDENT(5);
+		INDENT(5 + 9 * depth);
 		fprintf(fp, "[no tasks queued]\n");	
 	}
 }
-- 
1.7.1

>From 6df681dd0fb65615547602f24fa39823053a8376 Mon Sep 17 00:00:00 2001
From: zhangyanfei <zhangyanfei@xxxxxxxxxxxxxx>
Date: Thu, 30 Aug 2012 10:01:36 +0800
Subject: [PATCH] Fix rt not support group sched bug

Signed-off-by: zhangyanfei <zhangyanfei@xxxxxxxxxxxxxx>
---
 defs.h    |    2 ++
 symbols.c |    4 ++++
 task.c    |   49 +++++++++++++++++++++++++++++++++++++++----------
 3 files changed, 45 insertions(+), 10 deletions(-)

diff --git a/defs.h b/defs.h
index 4a8e2e3..4af670d 100755
--- a/defs.h
+++ b/defs.h
@@ -1785,6 +1785,7 @@ struct offset_table {                    /* stash of commonly-used offsets */
 	long log_level;
 	long log_flags_level;
 	long timekeeper_xtime_sec;
+	long sched_rt_entity_my_q;
 };
 
 struct size_table {         /* stash of commonly-used sizes */
@@ -1919,6 +1920,7 @@ struct size_table {         /* stash of commonly-used sizes */
 	long msg_queue;
 	long log;
 	long log_level;
+	long rt_rq;
 };
 
 struct array_table {
diff --git a/symbols.c b/symbols.c
index 2646ff8..bbadd5e 100755
--- a/symbols.c
+++ b/symbols.c
@@ -8812,6 +8812,8 @@ dump_offset_table(char *spec, ulong makestruct)
 		OFFSET(log_level));
 	fprintf(fp, "               log_flags_level: %ld\n",
 		OFFSET(log_flags_level));
+	fprintf(fp, "          sched_rt_entity_my_q: %ld\n",
+		OFFSET(sched_rt_entity_my_q));
 
 	fprintf(fp, "\n                    size_table:\n");
 	fprintf(fp, "                          page: %ld\n", SIZE(page));
@@ -9027,6 +9029,8 @@ dump_offset_table(char *spec, ulong makestruct)
 		SIZE(log));
 	fprintf(fp, "                     log_level: %ld\n",
 		SIZE(log_level));
+	fprintf(fp, "                         rt_rq: %ld\n",
+		SIZE(rt_rq));
 
         fprintf(fp, "\n                   array_table:\n");
 	/*
diff --git a/task.c b/task.c
index 6e4cfec..5b04e99 100755
--- a/task.c
+++ b/task.c
@@ -67,7 +67,7 @@ static void dump_task_runq_entry(struct task_context *);
 static int dump_tasks_in_cfs_rq(ulong);
 static void dump_on_rq_tasks(void);
 static void dump_CFS_runqueues(void);
-static void dump_RT_prio_array(ulong, char *);
+static void dump_RT_prio_array(int, ulong, char *);
 static void task_struct_member(struct task_context *,unsigned int, struct reference *);
 static void signal_reference(struct task_context *, ulong, struct reference *);
 static void do_sig_thread_group(ulong);
@@ -7552,6 +7552,7 @@ dump_CFS_runqueues(void)
 
 	if (!VALID_STRUCT(cfs_rq)) {
 		STRUCT_SIZE_INIT(cfs_rq, "cfs_rq");
+		STRUCT_SIZE_INIT(rt_rq, "rt_rq");
 		MEMBER_OFFSET_INIT(rq_rt, "rq", "rt");
 		MEMBER_OFFSET_INIT(rq_nr_running, "rq", "nr_running");
 		MEMBER_OFFSET_INIT(task_struct_se, "task_struct", "se");
@@ -7562,6 +7563,8 @@ dump_CFS_runqueues(void)
 			"cfs_rq");
 		MEMBER_OFFSET_INIT(sched_entity_my_q, "sched_entity", 
 			"my_q");
+		MEMBER_OFFSET_INIT(sched_rt_entity_my_q, "sched_rt_entity",
+			"my_q");
 		MEMBER_OFFSET_INIT(sched_entity_on_rq, "sched_entity", "on_rq");
 		MEMBER_OFFSET_INIT(cfs_rq_rb_leftmost, "cfs_rq", "rb_leftmost");
 		MEMBER_OFFSET_INIT(cfs_rq_nr_running, "cfs_rq", "nr_running");
@@ -7629,7 +7632,7 @@ dump_CFS_runqueues(void)
 				OFFSET(cfs_rq_tasks_timeline));
 		}
 
-		dump_RT_prio_array(runq + OFFSET(rq_rt) + OFFSET(rt_rq_active), 
+		dump_RT_prio_array(0, runq + OFFSET(rq_rt) + OFFSET(rt_rq_active),
 			&runqbuf[OFFSET(rq_rt) + OFFSET(rt_rq_active)]);
 
 		fprintf(fp, "  CFS RB_ROOT: %lx\n", (ulong)root);
@@ -7649,7 +7652,7 @@ dump_CFS_runqueues(void)
 }
 
 static void
-dump_RT_prio_array(ulong k_prio_array, char *u_prio_array)
+dump_RT_prio_array(int depth, ulong k_prio_array, char *u_prio_array)
 {
 	int i, c, tot, cnt, qheads;
 	ulong offset, kvaddr, uvaddr;
@@ -7657,8 +7660,11 @@ dump_RT_prio_array(ulong k_prio_array, char *u_prio_array)
         struct list_data list_data, *ld;
 	struct task_context *tc;
 	ulong *tlist;
+	ulong my_q, task_addr;
+	char *rt_rq_buf;
 
-	fprintf(fp, "  RT PRIO_ARRAY: %lx\n",  k_prio_array);
+	if (!depth)
+		fprintf(fp, "  RT PRIO_ARRAY: %lx\n",  k_prio_array);
 
         qheads = (i = ARRAY_LENGTH(rt_prio_array_queue)) ?
                 i : get_array_length("rt_prio_array.queue", NULL, SIZE(list_head));
@@ -7678,14 +7684,14 @@ dump_RT_prio_array(ulong k_prio_array, char *u_prio_array)
 		if ((list_head[0] == kvaddr) && (list_head[1] == kvaddr))
 			continue;
 
-		fprintf(fp, "     [%3d] ", i);
+		INDENT(5 + 8 * depth);
+		fprintf(fp, "[%3d] ", i);
 
 		BZERO(ld, sizeof(struct list_data));
 		ld->start = list_head[0];
 		if (VALID_MEMBER(task_struct_rt) &&
 		    VALID_MEMBER(sched_rt_entity_run_list))
-			ld->list_head_offset = OFFSET(task_struct_rt) + 
-				OFFSET(sched_rt_entity_run_list);
+			ld->list_head_offset = OFFSET(sched_rt_entity_run_list);
 		else
 			ld->list_head_offset = OFFSET(task_struct_run_list);
 		ld->end = kvaddr;
@@ -7695,10 +7701,33 @@ dump_RT_prio_array(ulong k_prio_array, char *u_prio_array)
 		tlist = (ulong *)GETBUF((cnt) * sizeof(ulong));
 		cnt = retrieve_list(tlist, cnt);
 		for (c = 0; c < cnt; c++) {
-			if (!(tc = task_to_context(tlist[c])))
+			task_addr = tlist[c];
+			if (VALID_MEMBER(sched_rt_entity_my_q)) {
+				readmem(tlist[c] + OFFSET(sched_rt_entity_my_q),
+					KVADDR, &my_q, sizeof(ulong), "my_q",
+					FAULT_ON_ERROR);
+				if (my_q) {
+					rt_rq_buf = GETBUF(SIZE(rt_rq));
+					readmem(my_q, KVADDR, rt_rq_buf,
+						SIZE(rt_rq), "rt_rq",
+						FAULT_ON_ERROR);
+					if (c)
+						INDENT(11 + 8 * depth);
+					fprintf(fp, "GROUP RT PRIO_ARRAY: %lx\n",
+						my_q + OFFSET(rt_rq_active));
+					tot++;
+					dump_RT_prio_array(depth + 1,
+						my_q + OFFSET(rt_rq_active),
+						&rt_rq_buf[OFFSET(rt_rq_active)]);
+					continue;
+				} else {
+					task_addr -= OFFSET(task_struct_rt);
+				}
+			}
+			if (!(tc = task_to_context(task_addr)))
 				continue;
 			if (c)
-				INDENT(11);
+				INDENT(11 + 8 * depth);
 			fprintf(fp, "PID: %-5ld  TASK: %lx  COMMAND: \"%s\"\n",
 				tc->pid, tc->task, tc->comm);
 			tot++;
@@ -7707,7 +7736,7 @@ dump_RT_prio_array(ulong k_prio_array, char *u_prio_array)
 	}
 
 	if (!tot) {
-		INDENT(5);
+		INDENT(5 + 8 * depth);
 		fprintf(fp, "[no tasks queued]\n");	
 	}
 }
-- 
1.7.1

--
Crash-utility mailing list
Crash-utility@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/crash-utility

[Index of Archives]     [Fedora Development]     [Fedora Desktop]     [Fedora SELinux]     [Yosemite News]     [KDE Users]     [Fedora Tools]

 

Powered by Linux