€•«UŒsphinx.addnodes”Œdocument”“”)”}”(Œ rawsource”Œ”Œchildren”]”(Œ translations”Œ LanguagesNode”“”)”}”(hhh]”(hŒ pending_xref”“”)”}”(hhh]”Œdocutils.nodes”ŒText”“”ŒChinese (Simplified)”…””}”Œparent”hsbaŒ attributes”}”(Œids”]”Œclasses”]”Œnames”]”Œdupnames”]”Œbackrefs”]”Œ refdomain”Œstd”Œreftype”Œdoc”Œ reftarget”Œ//translations/zh_CN/core-api/real-time/hardware”Œmodname”NŒ classname”NŒ refexplicit”ˆuŒtagname”hhh ubh)”}”(hhh]”hŒChinese (Traditional)”…””}”hh2sbah}”(h]”h ]”h"]”h$]”h&]”Œ refdomain”h)Œreftype”h+Œ reftarget”Œ//translations/zh_TW/core-api/real-time/hardware”Œmodname”NŒ classname”NŒ refexplicit”ˆuh1hhh ubh)”}”(hhh]”hŒItalian”…””}”hhFsbah}”(h]”h ]”h"]”h$]”h&]”Œ refdomain”h)Œreftype”h+Œ reftarget”Œ//translations/it_IT/core-api/real-time/hardware”Œmodname”NŒ classname”NŒ refexplicit”ˆuh1hhh ubh)”}”(hhh]”hŒJapanese”…””}”hhZsbah}”(h]”h ]”h"]”h$]”h&]”Œ refdomain”h)Œreftype”h+Œ reftarget”Œ//translations/ja_JP/core-api/real-time/hardware”Œmodname”NŒ classname”NŒ refexplicit”ˆuh1hhh ubh)”}”(hhh]”hŒKorean”…””}”hhnsbah}”(h]”h ]”h"]”h$]”h&]”Œ refdomain”h)Œreftype”h+Œ reftarget”Œ//translations/ko_KR/core-api/real-time/hardware”Œmodname”NŒ classname”NŒ refexplicit”ˆuh1hhh ubh)”}”(hhh]”hŒSpanish”…””}”hh‚sbah}”(h]”h ]”h"]”h$]”h&]”Œ refdomain”h)Œreftype”h+Œ reftarget”Œ//translations/sp_SP/core-api/real-time/hardware”Œmodname”NŒ classname”NŒ refexplicit”ˆuh1hhh ubeh}”(h]”h ]”h"]”h$]”h&]”Œcurrent_language”ŒEnglish”uh1h hhŒ _document”hŒsource”NŒline”NubhŒcomment”“”)”}”(hŒ SPDX-License-Identifier: GPL-2.0”h]”hŒ SPDX-License-Identifier: GPL-2.0”…””}”hh£sbah}”(h]”h ]”h"]”h$]”h&]”Œ xml:space”Œpreserve”uh1h¡hhhžhhŸŒI/var/lib/git/docbuild/linux/Documentation/core-api/real-time/hardware.rst”h KubhŒsection”“”)”}”(hhh]”(hŒtitle”“”)”}”(hŒConsidering hardware”h]”hŒConsidering hardware”…””}”(hh»hžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1h¹hh¶hžhhŸh³h KubhŒ field_list”“”)”}”(hhh]”hŒfield”“”)”}”(hhh]”(hŒ field_name”“”)”}”(hŒAuthor”h]”hŒAuthor”…””}”(hhÕhžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1hÓhhÐhŸh³h KubhŒ field_body”“”)”}”(hŒ2Sebastian Andrzej Siewior ”h]”hŒ paragraph”“”)”}”(hŒ1Sebastian Andrzej Siewior ”h]”(hŒSebastian Andrzej Siewior <”…””}”(hhëhžhhŸNh NubhŒ reference”“”)”}”(hŒbigeasy@linutronix.de”h]”hŒbigeasy@linutronix.de”…””}”(hhõhžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”Œrefuri”Œmailto:bigeasy@linutronix.de”uh1hóhhëubhŒ>”…””}”(hhëhžhhŸNh Nubeh}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h Khhåubah}”(h]”h ]”h"]”h$]”h&]”uh1hãhhÐubeh}”(h]”h ]”h"]”h$]”h&]”uh1hÎhŸh³h KhhËhžhubah}”(h]”h ]”h"]”h$]”h&]”uh1hÉhh¶hžhhŸh³h Kubhê)”}”(hXfThe way a workload is handled can be influenced by the hardware it runs on. Key components include the CPU, memory, and the buses that connect them. These resources are shared among all applications on the system. As a result, heavy utilization of one resource by a single application can affect the deterministic handling of workloads in other applications.”h]”hXfThe way a workload is handled can be influenced by the hardware it runs on. Key components include the CPU, memory, and the buses that connect them. These resources are shared among all applications on the system. As a result, heavy utilization of one resource by a single application can affect the deterministic handling of workloads in other applications.”…””}”(hj!hžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h K hh¶hžhubhê)”}”(hŒBelow is a brief overview.”h]”hŒBelow is a brief overview.”…””}”(hj/hžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h Khh¶hžhubhµ)”}”(hhh]”(hº)”}”(hŒSystem memory and cache”h]”hŒSystem memory and cache”…””}”(hj@hžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1h¹hj=hžhhŸh³h Kubhê)”}”(hXTMain memory and the associated caches are the most common shared resources among tasks in a system. One task can dominate the available caches, forcing another task to wait until a cache line is written back to main memory before it can proceed. The impact of this contention varies based on write patterns and the size of the caches available. Larger caches may reduce stalls because more lines can be buffered before being written back. Conversely, certain write patterns may trigger the cache controller to flush many lines at once, causing applications to stall until the operation completes.”h]”hXTMain memory and the associated caches are the most common shared resources among tasks in a system. One task can dominate the available caches, forcing another task to wait until a cache line is written back to main memory before it can proceed. The impact of this contention varies based on write patterns and the size of the caches available. Larger caches may reduce stalls because more lines can be buffered before being written back. Conversely, certain write patterns may trigger the cache controller to flush many lines at once, causing applications to stall until the operation completes.”…””}”(hjNhžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h Khj=hžhubhê)”}”(hX>This issue can be partly mitigated if applications do not share the same CPU cache. The kernel is aware of the cache topology and exports this information to user space. Tools such as **lstopo** from the Portable Hardware Locality (hwloc) project (https://www.open-mpi.org/projects/hwloc/) can visualize the hierarchy.”h]”(hŒ¸This issue can be partly mitigated if applications do not share the same CPU cache. The kernel is aware of the cache topology and exports this information to user space. Tools such as ”…””}”(hj\hžhhŸNh NubhŒstrong”“”)”}”(hŒ **lstopo**”h]”hŒlstopo”…””}”(hjfhžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1jdhj\ubhŒ6 from the Portable Hardware Locality (hwloc) project (”…””}”(hj\hžhhŸNh Nubhô)”}”(hŒ(https://www.open-mpi.org/projects/hwloc/”h]”hŒ(https://www.open-mpi.org/projects/hwloc/”…””}”(hjxhžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”Œrefuri”jzuh1hóhj\ubhŒ) can visualize the hierarchy.”…””}”(hj\hžhhŸNh Nubeh}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h Khj=hžhubhê)”}”(hXAvoiding shared L2 or L3 caches is not always possible. Even when cache sharing is minimized, bottlenecks can still occur when accessing system memory. Memory is used not only by the CPU but also by peripheral devices via DMA, such as graphics cards or network adapters.”h]”hXAvoiding shared L2 or L3 caches is not always possible. Even when cache sharing is minimized, bottlenecks can still occur when accessing system memory. Memory is used not only by the CPU but also by peripheral devices via DMA, such as graphics cards or network adapters.”…””}”(hj‘hžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h K"hj=hžhubhê)”}”(hX±In some cases, cache and memory bottlenecks can be controlled if the hardware provides the necessary support. On x86 systems, Intel offers Cache Allocation Technology (CAT), which enables cache partitioning among applications and provides control over the interconnect. AMD provides similar functionality under Platform Quality of Service (PQoS). On Arm64, the equivalent is Memory System Resource Partitioning and Monitoring (MPAM).”h]”hX±In some cases, cache and memory bottlenecks can be controlled if the hardware provides the necessary support. On x86 systems, Intel offers Cache Allocation Technology (CAT), which enables cache partitioning among applications and provides control over the interconnect. AMD provides similar functionality under Platform Quality of Service (PQoS). On Arm64, the equivalent is Memory System Resource Partitioning and Monitoring (MPAM).”…””}”(hjŸhžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h K'hj=hžhubhê)”}”(hŒ†These features can be configured through the Linux Resource Control interface. For details, see Documentation/filesystems/resctrl.rst.”h]”hŒ†These features can be configured through the Linux Resource Control interface. For details, see Documentation/filesystems/resctrl.rst.”…””}”(hj­hžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h K.hj=hžhubhê)”}”(hXPThe perf tool can be used to monitor cache behavior. It can analyze cache misses of an application and compare how they change under different workloads on a neighboring CPU. Even more powerful, the perf c2c tool can help identify cache-to-cache issues, where multiple CPU cores repeatedly access and modify data on the same cache line.”h]”hXPThe perf tool can be used to monitor cache behavior. It can analyze cache misses of an application and compare how they change under different workloads on a neighboring CPU. Even more powerful, the perf c2c tool can help identify cache-to-cache issues, where multiple CPU cores repeatedly access and modify data on the same cache line.”…””}”(hj»hžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h K1hj=hžhubeh}”(h]”Œsystem-memory-and-cache”ah ]”h"]”Œsystem memory and cache”ah$]”h&]”uh1h´hh¶hžhhŸh³h Kubhµ)”}”(hhh]”(hº)”}”(hŒHardware buses”h]”hŒHardware buses”…””}”(hjÔhžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1h¹hjÑhžhhŸh³h K8ubhê)”}”(hXXReal-time systems often need to access hardware directly to perform their work. Any latency in this process is undesirable, as it can affect the outcome of the task. For example, on an I/O bus, a changed output may not become immediately visible but instead appear with variable delay depending on the latency of the bus used for communication.”h]”hXXReal-time systems often need to access hardware directly to perform their work. Any latency in this process is undesirable, as it can affect the outcome of the task. For example, on an I/O bus, a changed output may not become immediately visible but instead appear with variable delay depending on the latency of the bus used for communication.”…””}”(hjâhžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h K:hjÑhžhubhê)”}”(hŒ»A bus such as PCI is relatively simple because register accesses are routed directly to the connected device. In the worst case, a read operation stalls the CPU until the device responds.”h]”hŒ»A bus such as PCI is relatively simple because register accesses are routed directly to the connected device. In the worst case, a read operation stalls the CPU until the device responds.”…””}”(hjðhžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h K@hjÑhžhubhê)”}”(hX›A bus such as USB is more complex, involving multiple layers. A register read or write is wrapped in a USB Request Block (URB), which is then sent by the USB host controller to the device. Timing and latency are influenced by the underlying USB bus. Requests cannot be sent immediately; they must align with the next frame boundary according to the endpoint type and the host controller's scheduling rules. This can introduce delays and additional latency. For example, a network device connected via USB may still deliver sufficient throughput, but the added latency when sending or receiving packets may fail to meet the requirements of certain real-time use cases.”h]”hXA bus such as USB is more complex, involving multiple layers. A register read or write is wrapped in a USB Request Block (URB), which is then sent by the USB host controller to the device. Timing and latency are influenced by the underlying USB bus. Requests cannot be sent immediately; they must align with the next frame boundary according to the endpoint type and the host controller’s scheduling rules. This can introduce delays and additional latency. For example, a network device connected via USB may still deliver sufficient throughput, but the added latency when sending or receiving packets may fail to meet the requirements of certain real-time use cases.”…””}”(hjþhžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h KDhjÑhžhubhê)”}”(hX³Additional restrictions on bus latency can arise from power management. For instance, PCIe with Active State Power Management (ASPM) enabled can suspend the link between the device and the host. While this behavior is beneficial for power savings, it delays device access and adds latency to responses. This issue is not limited to PCIe; internal buses within a System-on-Chip (SoC) can also be affected by power management mechanisms.”h]”hX³Additional restrictions on bus latency can arise from power management. For instance, PCIe with Active State Power Management (ASPM) enabled can suspend the link between the device and the host. While this behavior is beneficial for power savings, it delays device access and adds latency to responses. This issue is not limited to PCIe; internal buses within a System-on-Chip (SoC) can also be affected by power management mechanisms.”…””}”(hj hžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h KNhjÑhžhubeh}”(h]”Œhardware-buses”ah ]”h"]”Œhardware buses”ah$]”h&]”uh1h´hh¶hžhhŸh³h K8ubhµ)”}”(hhh]”(hº)”}”(hŒVirtualization”h]”hŒVirtualization”…””}”(hj%hžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1h¹hj"hžhhŸh³h KVubhê)”}”(hX§In a virtualized environment such as KVM, each guest CPU is represented as a thread on the host. If such a thread runs with real-time priority, the system should be tested to confirm it can sustain this behavior over extended periods. Because of its priority, the thread will not be preempted by lower-priority threads (such as SCHED_OTHER), which may then receive no CPU time. This can cause problems if a lower-priority thread is pinned to a CPU already occupied by a real-time task and unable to make progress. Even if a CPU has been isolated, the system may still (accidentally) start a per‑CPU thread on that CPU. Ensuring that a guest CPU goes idle is difficult, as it requires avoiding both task scheduling and interrupt handling. Furthermore, if the guest CPU does go idle but the guest system is booted with the option **idle=poll**, the guest CPU will never enter an idle state and will instead spin until an event arrives.”h]”(hX>In a virtualized environment such as KVM, each guest CPU is represented as a thread on the host. If such a thread runs with real-time priority, the system should be tested to confirm it can sustain this behavior over extended periods. Because of its priority, the thread will not be preempted by lower-priority threads (such as SCHED_OTHER), which may then receive no CPU time. This can cause problems if a lower-priority thread is pinned to a CPU already occupied by a real-time task and unable to make progress. Even if a CPU has been isolated, the system may still (accidentally) start a per‑CPU thread on that CPU. Ensuring that a guest CPU goes idle is difficult, as it requires avoiding both task scheduling and interrupt handling. Furthermore, if the guest CPU does go idle but the guest system is booted with the option ”…””}”(hj3hžhhŸNh Nubje)”}”(hŒ **idle=poll**”h]”hŒ idle=poll”…””}”(hj;hžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1jdhj3ubhŒ\, the guest CPU will never enter an idle state and will instead spin until an event arrives.”…””}”(hj3hžhhŸNh Nubeh}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h KXhj"hžhubhê)”}”(hXSDevice handling introduces additional considerations. Emulated PCI devices or VirtIO devices require a counterpart on the host to complete requests. This adds latency because the host must intercept and either process the request directly or schedule a thread for its completion. These delays can be avoided if the required PCI device is passed directly through to the guest. Some devices, such as networking or storage controllers, support the PCIe SR-IOV feature. SR-IOV allows a single PCIe device to be divided into multiple virtual functions, which can then be assigned to different guests.”h]”hXSDevice handling introduces additional considerations. Emulated PCI devices or VirtIO devices require a counterpart on the host to complete requests. This adds latency because the host must intercept and either process the request directly or schedule a thread for its completion. These delays can be avoided if the required PCI device is passed directly through to the guest. Some devices, such as networking or storage controllers, support the PCIe SR-IOV feature. SR-IOV allows a single PCIe device to be divided into multiple virtual functions, which can then be assigned to different guests.”…””}”(hjShžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h Kfhj"hžhubeh}”(h]”Œvirtualization”ah ]”h"]”Œvirtualization”ah$]”h&]”uh1h´hh¶hžhhŸh³h KVubhµ)”}”(hhh]”(hº)”}”(hŒ Networking”h]”hŒ Networking”…””}”(hjlhžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1h¹hjihžhhŸh³h Kpubhê)”}”(hŒùFor low-latency networking, the full networking stack may be undesirable, as it can introduce additional sources of delay. In this context, XDP can be used as a shortcut to bypass much of the stack while still relying on the kernel's network driver.”h]”hŒûFor low-latency networking, the full networking stack may be undesirable, as it can introduce additional sources of delay. In this context, XDP can be used as a shortcut to bypass much of the stack while still relying on the kernel’s network driver.”…””}”(hjzhžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h Krhjihžhubhê)”}”(hXaThe requirements are that the network driver must support XDP- preferably using an "skb pool" and that the application must use an XDP socket. Additional configuration may involve BPF filters, tuning networking queues, or configuring qdiscs for time-based transmission. These techniques are often applied in Time-Sensitive Networking (TSN) environments.”h]”hXeThe requirements are that the network driver must support XDP- preferably using an “skb pool†and that the application must use an XDP socket. Additional configuration may involve BPF filters, tuning networking queues, or configuring qdiscs for time-based transmission. These techniques are often applied in Time-Sensitive Networking (TSN) environments.”…””}”(hjˆhžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h Kwhjihžhubhê)”}”(hŒŽDocumenting all required steps exceeds the scope of this text. For detailed guidance, see the TSN documentation at https://tsn.readthedocs.io.”h]”(hŒsDocumenting all required steps exceeds the scope of this text. For detailed guidance, see the TSN documentation at ”…””}”(hj–hžhhŸNh Nubhô)”}”(hŒhttps://tsn.readthedocs.io”h]”hŒhttps://tsn.readthedocs.io”…””}”(hjžhžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”Œrefuri”j uh1hóhj–ubhŒ.”…””}”(hj–hžhhŸNh Nubeh}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h K}hjihžhubhê)”}”(hX3Another useful resource is the Linux Real-Time Communication Testbench https://github.com/Linutronix/RTC-Testbench. The goal of this project is to validate real-time network communication. It can be thought of as a "cyclictest" for networking and also serves as a starting point for application development.”h]”(hŒGAnother useful resource is the Linux Real-Time Communication Testbench ”…””}”(hj·hžhhŸNh Nubhô)”}”(hŒ+https://github.com/Linutronix/RTC-Testbench”h]”hŒ+https://github.com/Linutronix/RTC-Testbench”…””}”(hj¿hžhhŸNh Nubah}”(h]”h ]”h"]”h$]”h&]”Œrefuri”jÁuh1hóhj·ubhŒÅ. The goal of this project is to validate real-time network communication. It can be thought of as a “cyclictest†for networking and also serves as a starting point for application development.”…””}”(hj·hžhhŸNh Nubeh}”(h]”h ]”h"]”h$]”h&]”uh1héhŸh³h K€hjihžhubeh}”(h]”Œ networking”ah ]”h"]”Œ networking”ah$]”h&]”uh1h´hh¶hžhhŸh³h Kpubeh}”(h]”Œconsidering-hardware”ah ]”h"]”Œconsidering hardware”ah$]”h&]”uh1h´hhhžhhŸh³h Kubeh}”(h]”h ]”h"]”h$]”h&]”Œsource”h³uh1hŒcurrent_source”NŒ current_line”NŒsettings”Œdocutils.frontend”ŒValues”“”)”}”(h¹NŒ generator”NŒ datestamp”NŒ source_link”NŒ source_url”NŒ toc_backlinks”Œentry”Œfootnote_backlinks”KŒ sectnum_xform”KŒstrip_comments”NŒstrip_elements_with_classes”NŒ strip_classes”NŒ report_level”KŒ halt_level”KŒexit_status_level”KŒdebug”NŒwarning_stream”NŒ traceback”ˆŒinput_encoding”Œ utf-8-sig”Œinput_encoding_error_handler”Œstrict”Œoutput_encoding”Œutf-8”Œoutput_encoding_error_handler”j Œerror_encoding”Œutf-8”Œerror_encoding_error_handler”Œbackslashreplace”Œ language_code”Œen”Œrecord_dependencies”NŒconfig”NŒ id_prefix”hŒauto_id_prefix”Œid”Œ dump_settings”NŒdump_internals”NŒdump_transforms”NŒdump_pseudo_xml”NŒexpose_internals”NŒstrict_visitor”NŒ_disable_config”NŒ_source”h³Œ _destination”NŒ _config_files”]”Œ7/var/lib/git/docbuild/linux/Documentation/docutils.conf”aŒfile_insertion_enabled”ˆŒ raw_enabled”KŒline_length_limit”M'Œpep_references”NŒ pep_base_url”Œhttps://peps.python.org/”Œpep_file_url_template”Œpep-%04d”Œrfc_references”NŒ rfc_base_url”Œ&https://datatracker.ietf.org/doc/html/”Œ tab_width”KŒtrim_footnote_reference_space”‰Œsyntax_highlight”Œlong”Œ smart_quotes”ˆŒsmartquotes_locales”]”Œcharacter_level_inline_markup”‰Œdoctitle_xform”‰Œ docinfo_xform”KŒsectsubtitle_xform”‰Œ image_loading”Œlink”Œembed_stylesheet”‰Œcloak_email_addresses”ˆŒsection_self_link”‰Œenv”NubŒreporter”NŒindirect_targets”]”Œsubstitution_defs”}”Œsubstitution_names”}”Œrefnames”}”Œrefids”}”Œnameids”}”(jåjâjÎjËjjjfjcjÝjÚuŒ nametypes”}”(jå‰jΉj‰jf‰j݉uh}”(jâh¶jËj=jjÑjcj"jÚjiuŒ footnote_refs”}”Œ citation_refs”}”Œ autofootnotes”]”Œautofootnote_refs”]”Œsymbol_footnotes”]”Œsymbol_footnote_refs”]”Œ footnotes”]”Œ citations”]”Œautofootnote_start”KŒsymbol_footnote_start”KŒ id_counter”Œ collections”ŒCounter”“”}”…”R”Œparse_messages”]”Œtransform_messages”]”Œ transformer”NŒ include_log”]”Œ decoration”Nhžhub.