/fuchsia/zircon/kernel/include/kernel/ |
H A D | sched.h | 21 void sched_block(void) TA_REQ(thread_lock); 22 void sched_yield(void) TA_REQ(thread_lock); 23 void sched_preempt(void) TA_REQ(thread_lock); 24 void sched_reschedule(void) TA_REQ(thread_lock); 25 void sched_resched_internal(void) TA_REQ(thread_lock); 26 void sched_unblock_idle(thread_t* t) TA_REQ(thread_lock); 27 void sched_migrate(thread_t* t) TA_REQ(thread_lock); 31 void sched_inherit_priority(thread_t* t, int pri, bool* local_resched) TA_REQ(thread_lock); 35 void sched_change_priority(thread_t* t, int pri) TA_REQ(thread_lock); 39 bool sched_unblock(thread_t* t) __WARN_UNUSED_RESULT TA_REQ(thread_loc [all...] |
H A D | wait.h | 50 zx_status_t wait_queue_block(wait_queue_t*, zx_time_t deadline) TA_REQ(thread_lock); 58 uint signal_mask) TA_REQ(thread_lock); 63 int wait_queue_blocked_priority(wait_queue_t*) TA_REQ(thread_lock); 70 zx_status_t wait_queue_error) TA_REQ(thread_lock); 72 zx_status_t wait_queue_error) TA_REQ(thread_lock); 74 zx_status_t wait_queue_error) TA_REQ(thread_lock); 77 bool wait_queue_is_empty(wait_queue_t*) TA_REQ(thread_lock); 81 zx_status_t wait_queue_error) TA_REQ(thread_lock); 85 int old_prio) TA_REQ(thread_lock); 88 void wait_queue_validate_queue(wait_queue_t* wait) TA_REQ(thread_loc 104 zx_status_t Block(zx_time_t deadline) TA_REQ(thread_lock) { function in class:WaitQueue 108 zx_status_t BlockWithMask(zx_time_t deadline, uint signal_mask) TA_REQ(thread_lock) { function in class:WaitQueue 112 int BlockedPriority() TA_REQ(thread_lock) { function in class:WaitQueue 116 int WakeOne(bool reschedule, zx_status_t wait_queue_error) TA_REQ(thread_lock) { function in class:WaitQueue 120 int WakeAll(bool reschedule, zx_status_t wait_queue_error) TA_REQ(thread_lock) { function in class:WaitQueue 124 struct thread* DequeueOne(zx_status_t wait_queue_error) TA_REQ(thread_lock) { function in class:WaitQueue 128 bool IsEmpty() TA_REQ(thread_lock) { function in class:WaitQueue 132 static zx_status_t UnblockThread(struct thread* t, zx_status_t wait_queue_error) TA_REQ(thread_lock) { function in class:WaitQueue 136 static void PriorityChanged(struct thread* t, int old_prio) TA_REQ(thread_lock) { function in class:WaitQueue 140 void Validate() TA_REQ(thread_lock) { function in class:WaitQueue [all...] |
H A D | mp.h | 50 void mp_reschedule(cpu_mask_t mask, uint flags) TA_REQ(thread_lock); 110 static inline cpu_mask_t mp_get_idle_mask(void) TA_REQ(thread_lock) { function 114 static inline void mp_set_cpu_idle(cpu_num_t cpu) TA_REQ(thread_lock) { function 118 static inline void mp_set_cpu_busy(cpu_num_t cpu) TA_REQ(thread_lock) { function 122 static inline int mp_is_cpu_idle(cpu_num_t cpu) TA_REQ(thread_lock) { function 128 static inline void mp_set_cpu_realtime(cpu_num_t cpu) TA_REQ(thread_lock) { function 132 static inline void mp_set_cpu_non_realtime(cpu_num_t cpu) TA_REQ(thread_lock) { function 136 static inline cpu_mask_t mp_get_realtime_mask(void) TA_REQ(thread_lock) { function
|
H A D | dpc.h | 44 zx_status_t dpc_queue_thread_locked(dpc_t* dpc) TA_REQ(thread_lock);
|
/fuchsia/zircon/system/dev/block/fvm/ |
H A D | fvm-private.h | 153 zx_status_t WriteFvmLocked() TA_REQ(lock_); 156 size_t count) TA_REQ(lock_); 159 size_t count) TA_REQ(lock_); 161 zx_status_t FindFreeVPartEntryLocked(size_t* out) const TA_REQ(lock_); 162 zx_status_t FindFreeSliceLocked(size_t* out, size_t hint) const TA_REQ(lock_); 164 fvm_t* GetFvmLocked() const TA_REQ(lock_) { function in class:fvm::VPartitionManager 170 void FreePhysicalSlice(size_t pslice) TA_REQ(lock_); 174 void AllocatePhysicalSlice(size_t pslice, uint64_t vpart, uint64_t vslice) TA_REQ(lock_); 178 slice_entry_t* GetSliceEntryLocked(size_t index) const TA_REQ(lock_); 182 vpart_entry_t* GetVPartEntryLocked(size_t index) const TA_REQ(lock 184 size_t PrimaryOffsetLocked() const TA_REQ(lock_) { function in class:fvm::VPartitionManager 188 size_t BackupOffsetLocked() const TA_REQ(lock_) { function in class:fvm::VPartitionManager 231 auto ExtentBegin() TA_REQ(lock_) { function in class:fvm::VPartition 249 bool SliceCanFree(size_t vslice) const TA_REQ(lock_) { function in class:fvm::VPartition 264 void AddBlocksLocked(ssize_t nblocks) TA_REQ(lock_) { function in class:fvm::VPartition 270 void KillLocked() TA_REQ(lock_) { entry_index_ = 0; } function in class:fvm::VPartition 271 bool IsKilledLocked() TA_REQ(lock_) { return entry_index_ == 0; } function in class:fvm::VPartition [all...] |
/fuchsia/zircon/kernel/object/include/object/ |
H A D | socket_dispatcher.h | 74 void on_zero_handles_locked() TA_REQ(get_lock()); 75 void OnPeerZeroHandlesLocked() TA_REQ(get_lock()); 83 zx_status_t WriteSelfLocked(user_in_ptr<const void> src, size_t len, size_t* nwritten) TA_REQ(get_lock()); 84 zx_status_t WriteControlSelfLocked(user_in_ptr<const void> src, size_t len) TA_REQ(get_lock()); 85 zx_status_t UserSignalSelfLocked(uint32_t clear_mask, uint32_t set_mask) TA_REQ(get_lock()); 86 zx_status_t ShutdownOtherLocked(uint32_t how) TA_REQ(get_lock()); 87 zx_status_t ShareSelfLocked(HandleOwner h) TA_REQ(get_lock()); 89 bool is_full() const TA_REQ(get_lock()) { return data_.is_full(); } 90 bool is_empty() const TA_REQ(get_lock()) { return data_.is_empty(); }
|
H A D | event_pair_dispatcher.h | 29 void on_zero_handles_locked() TA_REQ(get_lock()); 30 void OnPeerZeroHandlesLocked() TA_REQ(get_lock());
|
H A D | futex_context.h | 55 void QueueNodesLocked(FutexNode* head) TA_REQ(lock_); 57 bool UnqueueNodeLocked(FutexNode* node) TA_REQ(lock_);
|
H A D | fifo_dispatcher.h | 37 void on_zero_handles_locked() TA_REQ(get_lock()); 38 void OnPeerZeroHandlesLocked() TA_REQ(get_lock()); 46 size_t* actual) TA_REQ(get_lock()); 47 zx_status_t UserSignalSelfLocked(uint32_t clear_mask, uint32_t set_mask) TA_REQ(get_lock());
|
H A D | timer_dispatcher.h | 38 void SetTimerLocked(bool cancel_first) TA_REQ(get_lock()); 39 bool CancelTimerLocked() TA_REQ(get_lock());
|
H A D | channel_dispatcher.h | 97 void on_zero_handles_locked() TA_REQ(get_lock()); 98 void OnPeerZeroHandlesLocked() TA_REQ(get_lock()); 108 void WriteSelf(fbl::unique_ptr<MessagePacket> msg) TA_REQ(get_lock()); 109 zx_status_t UserSignalSelf(uint32_t clear_mask, uint32_t set_mask) TA_REQ(get_lock());
|
H A D | bus_transaction_initiator_dispatcher.h | 70 void AddPmoLocked(PinnedMemoryTokenDispatcher* pmt) TA_REQ(get_lock()); 81 void PrintQuarantineWarningLocked() TA_REQ(get_lock());
|
/fuchsia/zircon/kernel/dev/iommu/intel/ |
H A D | iommu_impl.h | 65 void InvalidateIotlbDomainAllLocked(uint32_t domain_id) TA_REQ(lock_); 71 uint pages_pow2) TA_REQ(lock_); 93 void InvalidateContextCacheGlobalLocked() TA_REQ(lock_); 94 void InvalidateContextCacheDomainLocked(uint32_t domain_id) TA_REQ(lock_); 97 void InvalidateIotlbGlobalLocked() TA_REQ(lock_); 99 zx_status_t SetRootTablePointerLocked(paddr_t pa) TA_REQ(lock_); 100 zx_status_t SetTranslationEnableLocked(bool enabled, zx_time_t deadline) TA_REQ(lock_); 101 zx_status_t ConfigureFaultEventInterruptLocked() TA_REQ(lock_); 104 zx_status_t EnableBiosReservedMappingsLocked() TA_REQ(lock_); 106 void DisableFaultsLocked() TA_REQ(lock 120 volatile ds::RootTable* root_table() const TA_REQ(lock_) { function in class:intel_iommu::final [all...] |
/fuchsia/zircon/kernel/arch/arm64/include/arch/ |
H A D | aspace.h | 55 volatile pte_t* page_table) TA_REQ(lock_); 57 zx_status_t AllocPageTable(paddr_t* paddrp, uint page_size_shift) TA_REQ(lock_); 59 void FreePageTable(void* vaddr, paddr_t paddr, uint page_size_shift) TA_REQ(lock_); 64 volatile pte_t* page_table) TA_REQ(lock_); 68 volatile pte_t* page_table) TA_REQ(lock_); 72 volatile pte_t* page_table) TA_REQ(lock_); 80 uint page_size_shift) TA_REQ(lock_); 84 uint page_size_shift) TA_REQ(lock_); 88 uint top_index_shift, uint page_size_shift) TA_REQ(lock_); 89 zx_status_t QueryLocked(vaddr_t vaddr, paddr_t* paddr, uint* mmu_flags) TA_REQ(lock [all...] |
/fuchsia/zircon/kernel/arch/x86/page_tables/include/arch/x86/page_tables/ |
H A D | page_tables.h | 148 MappingCursor* new_cursor, ConsistencyManager* cm) TA_REQ(lock_); 151 MappingCursor* new_cursor, ConsistencyManager* cm) TA_REQ(lock_); 155 MappingCursor* new_cursor, ConsistencyManager* cm) TA_REQ(lock_); 158 MappingCursor* new_cursor, ConsistencyManager* cm) TA_REQ(lock_); 162 MappingCursor* new_cursor, ConsistencyManager* cm) TA_REQ(lock_); 165 ConsistencyManager* cm) TA_REQ(lock_); 170 volatile pt_entry_t** mapping) TA_REQ(lock_); 173 volatile pt_entry_t** mapping) TA_REQ(lock_); 176 volatile pt_entry_t* pte, ConsistencyManager* cm) TA_REQ(lock_); 180 bool was_terminal) TA_REQ(lock [all...] |
/fuchsia/zircon/system/dev/audio/intel-hda/controller/ |
H A D | intel-hda-stream.h | 77 void DeactivateLocked() TA_REQ(channel_lock_); 78 void EnsureStoppedLocked() TA_REQ(channel_lock_) { EnsureStopped(regs_); } function in class:audio::intel_hda::IntelHDAStream 84 TA_REQ(channel_lock_); 86 TA_REQ(channel_lock_); 87 zx_status_t ProcessStartLocked(const audio_proto::RingBufStartReq& req) TA_REQ(channel_lock_); 88 zx_status_t ProcessStopLocked(const audio_proto::RingBufStopReq& req) TA_REQ(channel_lock_); 91 void ReleaseRingBufferLocked() TA_REQ(channel_lock_);
|
H A D | intel-hda-controller.h | 84 void ReturnStreamLocked(fbl::RefPtr<IntelHDAStream>&& stream) TA_REQ (stream_pool_lock_); 85 uint8_t AllocateStreamTagLocked(bool input) TA_REQ (stream_pool_lock_); 86 void ReleaseStreamTagLocked (bool input, uint8_t tag_num) TA_REQ (stream_pool_lock_); 114 zx_status_t ResetCORBRdPtrLocked() TA_REQ(corb_lock_); 120 void ComputeCORBSpaceLocked() TA_REQ(corb_lock_); 121 void CommitCORBLocked() TA_REQ(corb_lock_); 122 void SendCodecCmdLocked(CodecCommand cmd) TA_REQ(corb_lock_); 126 zx_status_t HandleIrq() TA_REQ(default_domain_->token());
|
/fuchsia/zircon/system/dev/misc/tpm/ |
H A D | tpm.h | 88 uint8_t* resp, size_t max_len, size_t* actual) TA_REQ(lock_); 91 zx_status_t RequestLocalityLocked(Locality loc) TA_REQ(lock_); 93 zx_status_t WaitForLocalityLocked(Locality loc) TA_REQ(lock_); 95 zx_status_t ReleaseLocalityLocked(Locality loc) TA_REQ(lock_); 98 zx_status_t SendCmdLocked(Locality loc, const uint8_t* cmd, size_t len) TA_REQ(lock_); 102 size_t* actual) TA_REQ(lock_); 105 zx_status_t ShutdownLocked(uint16_t type) TA_REQ(lock_);
|
/fuchsia/zircon/system/dev/nand/nandpart/ |
H A D | aml-bad-block.h | 74 zx_status_t EraseBlock(uint32_t block) TA_REQ(lock_); 77 zx_status_t WritePages(uint32_t nand_page, uint32_t num_pages) TA_REQ(lock_); 80 zx_status_t ReadPages(uint32_t nand_page, uint32_t num_pages) TA_REQ(lock_); 83 zx_status_t GetNewBlock(void) TA_REQ(lock_); 86 zx_status_t WriteBadBlockTable(bool use_new_block) TA_REQ(lock_); 89 zx_status_t FindBadBlockTable(void) TA_REQ(lock_);
|
/fuchsia/zircon/system/dev/nand/skip-block/ |
H A D | skip-block.h | 58 zx_status_t GetBadBlockList(fbl::Array<uint32_t>* bad_block_list) TA_REQ(lock_); 63 zx_status_t GetPartitionInfo(skip_block_partition_info_t* info) const TA_REQ(lock_); 64 zx_status_t Read(const skip_block_rw_operation_t& info) TA_REQ(lock_); 65 zx_status_t Write(const skip_block_rw_operation_t& info, bool* bad_block_grown) TA_REQ(lock_);
|
/fuchsia/zircon/kernel/vm/include/vm/ |
H A D | vm_object.h | 47 virtual zx_status_t ResizeLocked(uint64_t size) TA_REQ(lock_) { return ZX_ERR_NOT_SUPPORTED; } function in class:VmObject 184 vm_page_t** page, paddr_t* pa) TA_REQ(lock_) { function in class:VmObject 191 void AddMappingLocked(VmMapping* r) TA_REQ(lock_); 192 void RemoveMappingLocked(VmMapping* r) TA_REQ(lock_); 203 void AddChildLocked(VmObject* r) TA_REQ(lock_); 204 void RemoveChildLocked(VmObject* r) TA_REQ(lock_); 235 void RangeChangeUpdateLocked(uint64_t offset, uint64_t len) TA_REQ(lock_);
|
H A D | vm_object_paged.h | 48 zx_status_t ResizeLocked(uint64_t size) override TA_REQ(lock_); 125 zx_status_t AddPageLocked(vm_page_t* p, uint64_t offset) TA_REQ(lock_); 130 zx_status_t PinLocked(uint64_t offset, uint64_t len) TA_REQ(lock_); 131 void UnpinLocked(uint64_t offset, uint64_t len) TA_REQ(lock_); 134 bool AnyPagesPinnedLocked(uint64_t offset, size_t len) TA_REQ(lock_); 141 zx_status_t SetParentOffsetLocked(uint64_t o) TA_REQ(lock_);
|
/fuchsia/zircon/system/private/zircon/ |
H A D | thread_annotations.h | 24 // TA_REQ(x) function requires that the caller hold the mutex |x| 43 #define TA_REQ(...) THREAD_ANNOTATION(requires_capability(__VA_ARGS__)) macro
|
/fuchsia/zircon/system/dev/bus/virtio/backends/ |
H A D | pci_legacy.cpp | 16 void PciLegacyBackend::IoReadLocked(uint16_t offset, uint8_t* val) TA_REQ(lock_) { function in namespace:virtio 20 void PciLegacyBackend::IoReadLocked(uint16_t offset, uint16_t* val) TA_REQ(lock_) { function in namespace:virtio 24 void PciLegacyBackend::IoReadLocked(uint16_t offset, uint32_t* val) TA_REQ(lock_) { function in namespace:virtio 28 void PciLegacyBackend::IoWriteLocked(uint16_t offset, uint8_t val) TA_REQ(lock_) { function in namespace:virtio 32 void PciLegacyBackend::IoWriteLocked(uint16_t offset, uint16_t val) TA_REQ(lock_) { function in namespace:virtio 36 void PciLegacyBackend::IoWriteLocked(uint16_t offset, uint32_t val) TA_REQ(lock_) { function in namespace:virtio
|
H A D | pci.h | 110 void CommonCfgCallbackLocked(const virtio_pci_cap_t& cap) TA_REQ(lock_); 111 void NotifyCfgCallbackLocked(const virtio_pci_cap_t& cap) TA_REQ(lock_); 112 void IsrCfgCallbackLocked(const virtio_pci_cap_t& cap) TA_REQ(lock_); 113 void DeviceCfgCallbackLocked(const virtio_pci_cap_t& cap) TA_REQ(lock_); 114 void PciCfgCallbackLocked(const virtio_pci_cap_t& cap) TA_REQ(lock_);
|