Index: sys/kern/kern_synch.c =================================================================== RCS file: /cvsroot/src/sys/kern/kern_synch.c,v retrieving revision 1.289 diff -p -r1.289 kern_synch.c *** sys/kern/kern_synch.c 13 May 2011 22:16:43 -0000 1.289 --- sys/kern/kern_synch.c 2 Jun 2011 06:38:51 -0000 *************** __KERNEL_RCSID(0, "$NetBSD: kern_synch.c *** 107,116 **** int dtrace_vtime_active=0; dtrace_vtime_switch_func_t dtrace_vtime_switch_func; ! static void sched_unsleep(struct lwp *, bool); ! static void sched_changepri(struct lwp *, pri_t); ! static void sched_lendpri(struct lwp *, pri_t); ! static void resched_cpu(struct lwp *); syncobj_t sleep_syncobj = { SOBJ_SLEEPQ_SORTED, --- 107,117 ---- int dtrace_vtime_active=0; dtrace_vtime_switch_func_t dtrace_vtime_switch_func; ! /*static*/ void sched_unsleep(struct lwp *, bool); ! /*static*/ void sched_changepri(struct lwp *, pri_t); ! /*static*/ void sched_lendpri(struct lwp *, pri_t); ! /*static*/ void resched_cpu(struct lwp *); ! /*static inline*/ lwp_t * nextlwp(struct cpu_info *, struct schedstate_percpu *); syncobj_t sleep_syncobj = { SOBJ_SLEEPQ_SORTED, *************** updatertime(lwp_t *l, const struct binti *** 534,540 **** /* * Select next LWP from the current CPU to run.. */ ! static inline lwp_t * nextlwp(struct cpu_info *ci, struct schedstate_percpu *spc) { lwp_t *newl; --- 535,541 ---- /* * Select next LWP from the current CPU to run.. */ ! /*static inline*/ lwp_t * nextlwp(struct cpu_info *ci, struct schedstate_percpu *spc) { lwp_t *newl; *************** suspendsched(void) *** 1094,1100 **** * interrupted: for example, if the sleep timed out. Because of this, * it's not a valid action for running or idle LWPs. */ ! static void sched_unsleep(struct lwp *l, bool cleanup) { --- 1095,1101 ---- * interrupted: for example, if the sleep timed out. Because of this, * it's not a valid action for running or idle LWPs. */ ! /*static*/ void sched_unsleep(struct lwp *l, bool cleanup) { *************** sched_unsleep(struct lwp *l, bool cleanu *** 1102,1108 **** panic("sched_unsleep"); } ! static void resched_cpu(struct lwp *l) { struct cpu_info *ci = l->l_cpu; --- 1103,1109 ---- panic("sched_unsleep"); } ! /*static*/ void resched_cpu(struct lwp *l) { struct cpu_info *ci = l->l_cpu; *************** resched_cpu(struct lwp *l) *** 1112,1118 **** cpu_need_resched(ci, 0); } ! static void sched_changepri(struct lwp *l, pri_t pri) { --- 1113,1119 ---- cpu_need_resched(ci, 0); } ! /*static*/ void sched_changepri(struct lwp *l, pri_t pri) { *************** sched_changepri(struct lwp *l, pri_t pri *** 1129,1135 **** resched_cpu(l); } ! static void sched_lendpri(struct lwp *l, pri_t pri) { --- 1130,1136 ---- resched_cpu(l); } ! /*static*/ void sched_lendpri(struct lwp *l, pri_t pri) {