base synchronization/lock primitive support More...

Files

file  lock_core.h
 

Macros

#define lock_owner_id_t   int8_t
 type to use to store the 'owner' of a lock. More...
 
#define LOCK_INVALID_OWNER_ID   ((lock_owner_id_t)-1)
 marker value to use for a lock_owner_id_t which does not refer to any valid owner
 
#define lock_get_caller_owner_id()   ((lock_owner_id_t)get_core_num())
 return the owner id for the caller More...
 
#define lock_internal_spin_unlock_with_wait(lock, save)   spin_unlock((lock)->spin_lock, save), __wfe()
 Atomically unlock the lock's spin lock, and wait for a notification. More...
 
#define lock_internal_spin_unlock_with_notify(lock, save)   spin_unlock((lock)->spin_lock, save), __sev()
 Atomically unlock the lock's spin lock, and send a notification. More...
 
#define lock_internal_spin_unlock_with_best_effort_wait_or_timeout(lock, save, until)
 Atomically unlock the lock's spin lock, and wait for a notification or a timeout. More...
 
#define sync_internal_yield_until_before(until)   ((void)0)
 yield to other processing until some time before the requested time More...
 

Functions

void lock_init (lock_core_t *core, uint lock_num)
 Initialise a lock structure. More...
 

Detailed Description

base synchronization/lock primitive support

Most of the pico_sync locking primitives contain a lock_core_t structure member. This currently just holds a spin lock which is used only to protect the contents of the rest of the structure as part of implementing the synchronization primitive. As such, the spin_lock member of lock core is never still held on return from any function for the primitive.

critical_section is an exceptional case in that it does not have a lock_core_t and simply wraps a spin lock, providing methods to lock and unlock said spin lock.

lock_core based structures work by locking the spin lock, checking state, and then deciding whether they additionally need to block or notify when the spin lock is released. In the blocking case, they will wake up again in the future, and try the process again.

By default the SDK just uses the processors' events via SEV and WEV for notification and blocking as these are sufficient for cross core, and notification from interrupt handlers. However macros are defined in this file that abstract the wait and notify mechanisms to allow the SDK locking functions to effectively be used within an RTOS or other environment.

When implementing an RTOS, it is desirable for the SDK synchronization primitives that wait, to block the calling task (and immediately yield), and those that notify, to wake a blocked task which isn't on processor. At least the wait macro implementation needs to be atomic with the protecting spin_lock unlock from the callers point of view; i.e. the task should unlock the spin lock when it starts its wait. Such implementation is up to the RTOS integration, however the macros are defined such that such operations are always combined into a single call (so they can be perfomed atomically) even though the default implementation does not need this, as a WFE which starts following the corresponding SEV is not missed.

Macro Definition Documentation

◆ lock_get_caller_owner_id

#define lock_get_caller_owner_id ( )    ((lock_owner_id_t)get_core_num())

return the owner id for the caller

By default this returns the calling core number, but may be overridden (e.g. to return an RTOS task id)

◆ lock_internal_spin_unlock_with_best_effort_wait_or_timeout

#define lock_internal_spin_unlock_with_best_effort_wait_or_timeout (   lock,
  save,
  until 
)
Value:
({ \
spin_unlock((lock)->spin_lock, save); \
best_effort_wfe_or_timeout(until); \
})

Atomically unlock the lock's spin lock, and wait for a notification or a timeout.

Atomic here refers to the fact that it should not be possible for a concurrent lock_internal_spin_unlock_with_notify to insert itself between the spin unlock and this wait in a way that the wait does not see the notification (i.e. causing a missed notification). In other words this method should always wake up in response to a lock_internal_spin_unlock_with_notify for the same lock, which completes after this call starts.

In an ideal implementation, this method would return exactly after the corresponding lock_internal_spin_unlock_with_notify has subsequently been called on the same lock instance or the timeout has been reached, however this method is free to return at any point before that; this macro is always used in a loop which locks the spin lock, checks the internal locking primitive state and then waits again if the calling thread should not proceed.

By default this simply unlocks the spin lock, and then calls best_effort_wfe_or_timeout but may be overridden (e.g. to actually block the RTOS task with a timeout).

Parameters
lockthe lock_core for the primitive which needs to block
savethe uint32_t value that should be passed to spin_unlock when the spin lock is unlocked. (i.e. the PRIMASK state when the spin lock was acquire)
untilthe absolute_time_t value
Returns
true if the timeout has been reached

◆ lock_internal_spin_unlock_with_notify

#define lock_internal_spin_unlock_with_notify (   lock,
  save 
)    spin_unlock((lock)->spin_lock, save), __sev()

Atomically unlock the lock's spin lock, and send a notification.

Atomic here refers to the fact that it should not be possible for this notification to happen during a lock_internal_spin_unlock_with_wait in a way that that wait does not see the notification (i.e. causing a missed notification). In other words this method should always wake up any lock_internal_spin_unlock_with_wait which started before this call completes.

In an ideal implementation, this method would wake up only the corresponding lock_internal_spin_unlock_with_wait that has been called on the same lock instance, however it is free to wake up any of them, as they will check their condition and then re-wait if necessary/

By default this macro simply unlocks the spin lock, and then performs a SEV, but may be overridden (e.g. to actually un-block RTOS task(s)).

Parameters
lockthe lock_core for the primitive which needs to block
savethe uint32_t value that should be passed to spin_unlock when the spin lock is unlocked. (i.e. the PRIMASK state when the spin lock was acquire)

◆ lock_internal_spin_unlock_with_wait

#define lock_internal_spin_unlock_with_wait (   lock,
  save 
)    spin_unlock((lock)->spin_lock, save), __wfe()

Atomically unlock the lock's spin lock, and wait for a notification.

Atomic here refers to the fact that it should not be possible for a concurrent lock_internal_spin_unlock_with_notify to insert itself between the spin unlock and this wait in a way that the wait does not see the notification (i.e. causing a missed notification). In other words this method should always wake up in response to a lock_internal_spin_unlock_with_notify for the same lock, which completes after this call starts.

In an ideal implementation, this method would return exactly after the corresponding lock_internal_spin_unlock_with_notify has subsequently been called on the same lock instance, however this method is free to return at any point before that; this macro is always used in a loop which locks the spin lock, checks the internal locking primitive state and then waits again if the calling thread should not proceed.

By default this macro simply unlocks the spin lock, and then performs a WFE, but may be overridden (e.g. to actually block the RTOS task).

Parameters
lockthe lock_core for the primitive which needs to block
savethe uint32_t value that should be passed to spin_unlock when the spin lock is unlocked. (i.e. the PRIMASK state when the spin lock was acquire

◆ lock_owner_id_t

#define lock_owner_id_t   int8_t

type to use to store the 'owner' of a lock.

By default this is int8_t as it only needs to store the core number or -1, however it may be overridden if a larger type is required (e.g. for an RTOS task id)

◆ sync_internal_yield_until_before

#define sync_internal_yield_until_before (   until)    ((void)0)

yield to other processing until some time before the requested time

This method is provided for cases where the caller has no useful work to do until the specified time.

By default this method does nothing, however it can be overridden (for example by an RTOS which is able to block the current task until the scheduler tick before the given time)

Parameters
untilthe absolute_time_t value

Function Documentation

◆ lock_init()

void lock_init ( lock_core_t core,
uint  lock_num 
)

Initialise a lock structure.

Inititalize a lock structure, providing the spin lock number to use for protecting internal state.

Parameters
corePointer to the lock_core to initialize
lock_numSpin lock number to use for the lock. As the spin lock is only used internally to the locking primitive method implementations, this does not need to be globally unique, however could suffer contention