@@ -218,7 +218,7 @@ static inline spinlock_t up_testset(FAR volatile spinlock_t *lock)
218
218
#define spin_lock_init (l ) do { *(l) = SP_UNLOCKED; } while (0)
219
219
220
220
/****************************************************************************
221
- * Name: spin_lock_wo_note
221
+ * Name: spin_lock_notrace
222
222
*
223
223
* Description:
224
224
* If this CPU does not already hold the spinlock, then loop until the
@@ -240,7 +240,7 @@ static inline spinlock_t up_testset(FAR volatile spinlock_t *lock)
240
240
****************************************************************************/
241
241
242
242
#ifdef CONFIG_SPINLOCK
243
- static inline_function void spin_lock_wo_note (FAR volatile spinlock_t * lock )
243
+ static inline_function void spin_lock_notrace (FAR volatile spinlock_t * lock )
244
244
{
245
245
#ifdef CONFIG_TICKET_SPINLOCK
246
246
unsigned short ticket =
@@ -290,7 +290,7 @@ static inline_function void spin_lock(FAR volatile spinlock_t *lock)
290
290
291
291
/* Lock without trace note */
292
292
293
- spin_lock_wo_note (lock );
293
+ spin_lock_notrace (lock );
294
294
295
295
/* Notify that we have the spinlock */
296
296
@@ -299,7 +299,7 @@ static inline_function void spin_lock(FAR volatile spinlock_t *lock)
299
299
#endif /* CONFIG_SPINLOCK */
300
300
301
301
/****************************************************************************
302
- * Name: spin_trylock_wo_note
302
+ * Name: spin_trylock_notrace
303
303
*
304
304
* Description:
305
305
* Try once to lock the spinlock. Do not wait if the spinlock is already
@@ -322,7 +322,7 @@ static inline_function void spin_lock(FAR volatile spinlock_t *lock)
322
322
323
323
#ifdef CONFIG_SPINLOCK
324
324
static inline_function bool
325
- spin_trylock_wo_note (FAR volatile spinlock_t * lock )
325
+ spin_trylock_notrace (FAR volatile spinlock_t * lock )
326
326
{
327
327
#ifdef CONFIG_TICKET_SPINLOCK
328
328
unsigned short ticket =
@@ -387,7 +387,7 @@ static inline_function bool spin_trylock(FAR volatile spinlock_t *lock)
387
387
388
388
/* Try lock without trace note */
389
389
390
- locked = spin_trylock_wo_note (lock );
390
+ locked = spin_trylock_notrace (lock );
391
391
if (locked )
392
392
{
393
393
/* Notify that we have the spinlock */
@@ -406,7 +406,7 @@ static inline_function bool spin_trylock(FAR volatile spinlock_t *lock)
406
406
#endif /* CONFIG_SPINLOCK */
407
407
408
408
/****************************************************************************
409
- * Name: spin_unlock_wo_note
409
+ * Name: spin_unlock_notrace
410
410
*
411
411
* Description:
412
412
* Release one count on a non-reentrant spinlock.
@@ -427,7 +427,7 @@ static inline_function bool spin_trylock(FAR volatile spinlock_t *lock)
427
427
428
428
#ifdef CONFIG_SPINLOCK
429
429
static inline_function void
430
- spin_unlock_wo_note (FAR volatile spinlock_t * lock )
430
+ spin_unlock_notrace (FAR volatile spinlock_t * lock )
431
431
{
432
432
SP_DMB ();
433
433
#ifdef CONFIG_TICKET_SPINLOCK
@@ -463,7 +463,7 @@ static inline_function void spin_unlock(FAR volatile spinlock_t *lock)
463
463
{
464
464
/* Unlock without trace note */
465
465
466
- spin_unlock_wo_note (lock );
466
+ spin_unlock_notrace (lock );
467
467
468
468
/* Notify that we are unlocking the spinlock */
469
469
@@ -496,7 +496,7 @@ static inline_function void spin_unlock(FAR volatile spinlock_t *lock)
496
496
#endif
497
497
498
498
/****************************************************************************
499
- * Name: spin_lock_irqsave_wo_note
499
+ * Name: spin_lock_irqsave_notrace
500
500
*
501
501
* Description:
502
502
* This function is no trace version of spin_lock_irqsave()
@@ -505,7 +505,7 @@ static inline_function void spin_unlock(FAR volatile spinlock_t *lock)
505
505
506
506
#ifdef CONFIG_SPINLOCK
507
507
static inline_function
508
- irqstate_t spin_lock_irqsave_wo_note (FAR volatile spinlock_t * lock )
508
+ irqstate_t spin_lock_irqsave_notrace (FAR volatile spinlock_t * lock )
509
509
{
510
510
irqstate_t ret ;
511
511
ret = up_irq_save ();
@@ -515,21 +515,21 @@ irqstate_t spin_lock_irqsave_wo_note(FAR volatile spinlock_t *lock)
515
515
int me = this_cpu ();
516
516
if (0 == g_irq_spin_count [me ])
517
517
{
518
- spin_lock_wo_note (& g_irq_spin );
518
+ spin_lock_notrace (& g_irq_spin );
519
519
}
520
520
521
521
g_irq_spin_count [me ]++ ;
522
522
DEBUGASSERT (0 != g_irq_spin_count [me ]);
523
523
}
524
524
else
525
525
{
526
- spin_lock_wo_note (lock );
526
+ spin_lock_notrace (lock );
527
527
}
528
528
529
529
return ret ;
530
530
}
531
531
#else
532
- # define spin_lock_irqsave_wo_note (l ) ((void)(l), up_irq_save())
532
+ # define spin_lock_irqsave_notrace (l ) ((void)(l), up_irq_save())
533
533
#endif
534
534
535
535
/****************************************************************************
@@ -577,7 +577,7 @@ irqstate_t spin_lock_irqsave(FAR volatile spinlock_t *lock)
577
577
578
578
/* Lock without trace note */
579
579
580
- flags = spin_lock_irqsave_wo_note (lock );
580
+ flags = spin_lock_irqsave_notrace (lock );
581
581
582
582
/* Notify that we have the spinlock */
583
583
@@ -590,7 +590,7 @@ irqstate_t spin_lock_irqsave(FAR volatile spinlock_t *lock)
590
590
#endif
591
591
592
592
/****************************************************************************
593
- * Name: spin_trylock_irqsave_wo_note
593
+ * Name: spin_trylock_irqsave_notrace
594
594
*
595
595
* Description:
596
596
* Try once to lock the spinlock. Do not wait if the spinlock is already
@@ -613,14 +613,14 @@ irqstate_t spin_lock_irqsave(FAR volatile spinlock_t *lock)
613
613
****************************************************************************/
614
614
615
615
#ifdef CONFIG_SPINLOCK
616
- # define spin_trylock_irqsave_wo_note (l , f ) \
616
+ # define spin_trylock_irqsave_notrace (l , f ) \
617
617
({ \
618
618
f = up_irq_save(); \
619
- spin_trylock_wo_note (l) ? \
619
+ spin_trylock_notrace (l) ? \
620
620
true : ({ up_irq_restore(f); false; }); \
621
621
})
622
622
#else
623
- # define spin_trylock_irqsave_wo_note (l , f ) \
623
+ # define spin_trylock_irqsave_notrace (l , f ) \
624
624
({ \
625
625
(void)(l); \
626
626
f = up_irq_save(); \
@@ -665,7 +665,7 @@ irqstate_t spin_lock_irqsave(FAR volatile spinlock_t *lock)
665
665
#endif /* CONFIG_SPINLOCK */
666
666
667
667
/****************************************************************************
668
- * Name: spin_unlock_irqrestore_wo_note
668
+ * Name: spin_unlock_irqrestore_notrace
669
669
*
670
670
* Description:
671
671
* This function is no trace version of spin_unlock_irqrestore()
@@ -674,7 +674,7 @@ irqstate_t spin_lock_irqsave(FAR volatile spinlock_t *lock)
674
674
675
675
#ifdef CONFIG_SPINLOCK
676
676
static inline_function
677
- void spin_unlock_irqrestore_wo_note (FAR volatile spinlock_t * lock ,
677
+ void spin_unlock_irqrestore_notrace (FAR volatile spinlock_t * lock ,
678
678
irqstate_t flags )
679
679
{
680
680
if (NULL == lock )
@@ -685,18 +685,18 @@ void spin_unlock_irqrestore_wo_note(FAR volatile spinlock_t *lock,
685
685
686
686
if (0 == g_irq_spin_count [me ])
687
687
{
688
- spin_unlock_wo_note (& g_irq_spin );
688
+ spin_unlock_notrace (& g_irq_spin );
689
689
}
690
690
}
691
691
else
692
692
{
693
- spin_unlock_wo_note (lock );
693
+ spin_unlock_notrace (lock );
694
694
}
695
695
696
696
up_irq_restore (flags );
697
697
}
698
698
#else
699
- # define spin_unlock_irqrestore_wo_note (l , f ) ((void)(l), up_irq_restore(f))
699
+ # define spin_unlock_irqrestore_notrace (l , f ) ((void)(l), up_irq_restore(f))
700
700
#endif
701
701
702
702
/****************************************************************************
@@ -735,7 +735,7 @@ void spin_unlock_irqrestore(FAR volatile spinlock_t *lock,
735
735
{
736
736
/* Unlock without trace note */
737
737
738
- spin_unlock_irqrestore_wo_note (lock , flags );
738
+ spin_unlock_irqrestore_notrace (lock , flags );
739
739
740
740
/* Notify that we are unlocking the spinlock */
741
741
0 commit comments