You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
125 lines
12 KiB
125 lines
12 KiB
\chapter{Technical Background}
|
|
\label{chap:state}
|
|
|
|
% Hier werden zwei wesentliche Aufgaben erledigt:
|
|
|
|
% 1. Der Leser muß alles beigebracht bekommen, was er zum Verständnis
|
|
% der späteren Kapitel braucht. Insbesondere sind in unserem Fach die
|
|
% Systemvoraussetzungen zu klären, die man später benutzt. Zulässig ist
|
|
% auch, daß man hier auf Tutorials oder Ähnliches verweist, die hier auf
|
|
% dem Netz zugänglich sind.
|
|
|
|
% 2. Es muß klar werden, was anderswo zu diesem Problem gearbeitet
|
|
% wird. Insbesondere sollen natürlich die Lücken der anderen klar
|
|
% werden. Warum ist die eigene Arbeit, der eigene Ansatz wichtig, um
|
|
% hier den Stand der Technik weiterzubringen? Dieses Kapitel wird von
|
|
% vielen Lesern übergangen (nicht aber vom Gutachter ;-), auch später
|
|
% bei Veröffentlichungen ist "Related Work" eine wichtige Sache.
|
|
|
|
% Viele Leser stellen dann später fest, daß sie einige der Grundlagen
|
|
% doch brauchen und blättern zurück. Deshalb ist es gut,
|
|
% Rückwärtsverweise in späteren Kapiteln zu haben, und zwar so, daß man
|
|
% die Abschnitte, auf die verwiesen wird, auch für sich lesen
|
|
% kann. Diese Kapitel kann relativ lang werden, je größer der Kontext
|
|
% der Arbeit, desto länger. Es lohnt sich auch! Den Text kann man unter
|
|
% Umständen wiederverwenden, indem man ihn als "Tutorial" zu einem
|
|
% Gebiet auch dem Netz zugänglich macht.
|
|
|
|
% Dadurch gewinnt man manchmal wertvolle Hinweise von Kollegen. Dieses
|
|
% Kapitel wird in der Regel zuerst geschrieben und ist das Einfachste
|
|
% (oder das Schwerste weil erste).
|
|
|
|
\todo{write introductory paragraph}
|
|
|
|
\section{High Bandwidth Memory}
|
|
|
|
\glsentrylong{hbm} is a novel memory technology promising an increase in peak bandwidth. It is composed of stacked DRAM dies \cite[p. 1]{hbm-arch-paper} and is slowly being integrated into server processors, notably the Intel® Xeon® Max Series \cite{intel:xeonmaxbrief}. \gls{hbm} on these systems may be configured in different memory modes, most notably, HBM Flat Mode and HBM Cache Mode \cite{intel:xeonmaxbrief}. The former gives applications direct control, requiring code changes while the latter utilizes the \gls{hbm} as cache for the systems DDR based main memory \cite{intel:xeonmaxbrief}. \par
|
|
|
|
\section{Query Driven Prefetching}
|
|
|
|
\todo{write this section}
|
|
|
|
\section{Intel Data Streaming Accelerator}
|
|
|
|
\blockquote{Intel \gls{dsa} is a high-performance data copy and transformation accelerator that will be integrated in future Intel® processors, targeted for optimizing streaming data movement and transformation operations common with applications for high-performance storage, networking, persistent memory, and various data processing applications. \cite[15]{intel:dsaspec}}
|
|
|
|
Introduced with the 4th generation of Intel Xeon Scalable Processors \cite{intel:xeonbrief}, the \gls{dsa} promises to alleviate the CPU from \enquote{common storage functions and operations such as data integrity checks and deduplication} \cite{intel:xeonbrief}. This chapter will give an overview of the architecture, software and the interaction of these two components. The reader will be familiarized with the setup and equipped with the knowledge to configure the system for a specific use case.
|
|
|
|
\todo{consider adding projected use cases as in the architecture specification here}
|
|
To be able to optimally utilize the Hardware, knowledge of its workings is required to make educated decisions. Therefore, this section describes both the workings of the \gls{dsa} engine itself and the view that is presented through software interfaces. All statements are based on Chapter 3 of the Architecture Specification by Intel \cite{intel:dsaspec}. \par
|
|
|
|
\subsection{Hardware Architecture} \label{subsection:dsa-hwarch}
|
|
|
|
\begin{figure}[H]
|
|
\centering
|
|
\includegraphics[width=0.9\textwidth]{images/dsa-internal-block-diagram.png}
|
|
\caption{\\ \acrshort{dsa} Internal Archtiecture Block Diagramm \\ Taken from Figure 1a of \cite{intel:analysis}}
|
|
\label{fig:dsa-internal-block}
|
|
\end{figure}
|
|
|
|
The accelerator is directly integrated into the Processor and attaches via the I/O fabric interface over which all communication is conducted. Over this interface, it is accessible as a PCIe device. Configuration therefore is done through memory-mapped registers set in the devices \gls{bar}. Through these, the devices' layout is defined and memory pages for work submission are set. In a system with multiple processing nodes, there may also be one \gls{dsa} per node.\par
|
|
|
|
To satisfy different use cases, as already mentioned, the layout of the \gls{dsa} may be software-defined. The structure is made up of three components, namely \gls{dsa:wq}s, \gls{dsa:engine}s and \gls{dsa:group}s. \gls{dsa:wq}s provide the means to submit tasks to the device and will be described in more detail shortly. An \gls{dsa:engine} is the processing-block that connects to memory and performs the described task. Using \gls{dsa:group}s, \gls{dsa:engine}s and \gls{dsa:wq}s are tied together. This means, that tasks from one \gls{dsa:wq} may be processed from multiple \gls{dsa:engine}s and that vice-versa, depending on the configuration. This flexibility is achieved through the Group Arbiter which connects the two components and acts according to the setup. \par
|
|
|
|
A \gls{dsa:wq} is accessible through so-called portals, which are mapped memory regions. Submission of work is done by writing a descriptor to one of these portals. A descriptor is 64 bytes in size and may contain one specific task (task descriptor) or the location of a task array in memory (batch descriptor). Through these portals, the submitted descriptor reaches a queue of which there are two types with different submission methods and use cases. The \gls{dsa:swq} is intended to provide synchronized access to multiple processes and each group may only have one attached. A \gls{pcie-dmr}, which guarantees implicit synchronization, is generated via \gls{x86:enqcmd} and communicates with the device before writing. This results in higher submission cost, compared to the \gls{dsa:dwq} to which a descriptor is submitted via \gls{x86:movdir64b}. The \gls{dsa:dwq} is therefore more performant but may require access control mechanisms and may only be accessed by one process at a time. \par
|
|
|
|
To handle the different descriptors, each \gls{dsa:engine} has two internal execution paths. One for a task and the other for a batch descriptor. Processing a task descriptor is straightforward, as all information required to complete the operation are contained within. For a batch, the \gls{dsa} first reads the batch descriptor, then fetches all task descriptors for the batch from memory and processes them. An \gls{dsa:engine} can also trigger a page fault when trying to access an unloaded page and wait on its completion, if configured to do so. Otherwise, an error will be generated in this scenario. \par
|
|
|
|
Ordering of operations is only guaranteed for a configuration with one \gls{dsa:wq} and one \gls{dsa:engine} in a \gls{dsa:group} when submitting exclusively batch or task descriptors but no mixture. Even then, only write-ordering is guaranteed, meaning that \enquote{reads by a subsequent descriptor can pass writes from a previous descriptor} \cite[30]{intel:dsaspec}. A different issue arises, should an operation fail: the \gls{dsa} will continue to process the following descriptors. Care must therefore be taken with read-after-write scenarios, either by waiting for a successful completion before submitting the dependant, inserting a drain descriptor for tasks or setting the fence flag for a batch. The latter two methods tell the processing engine that all writes must be committed and, in case of the fence in a batch, abort on previous error. \par
|
|
|
|
An important aspect of modern computer systems is the separation of address spaces through virtual memory. The \gls{dsa} must therefore handle address translation, as a process submitting a task will not know the physical location in memory which causes the descriptor to contain virtual values. For this, the \gls{dsa:engine} communicates with the \gls{iommu} and \gls{atc} to perform this operation. For this, knowledge about the submitting processes is required, and therefore each task descriptor has a field for the \gls{x86:pasid} which is filled by the \gls{x86:enqcmd} instruction for a \gls{dsa:swq} or set statically after a process is attached to a \gls{dsa:dwq}. \par
|
|
|
|
The completion of a descriptor may be signalled through a completion record and interrupt, if configured so. For this, the \gls{dsa} \enquote{provides two types of interrupt message storage: (1) an MSI-X table, enumerated through the MSI-X capability; and (2) a device-specific Interrupt Message Storage (IMS) table} \cite[27]{intel:dsaspec}. \par
|
|
|
|
\subsection{Software View}
|
|
|
|
\begin{figure}[H]
|
|
\centering
|
|
\includegraphics[width=0.5\textwidth]{images/dsa-software-architecture.png}
|
|
\caption{\\ \acrshort{dsa} Software View Block Diagramm \\ Taken from Figure 1a of \cite{intel:analysis}}
|
|
\label{fig:dsa-software-arch}
|
|
\end{figure}
|
|
|
|
Due to efforts by intel programmers, since Linux Kernel 5.10 \cite[Installation Instructinos]{intel:dmldoc}, there exists a driver for the \gls{dsa} \cite{intel:idxd-driver-repo} which has no counterpart in the Windows OS-Family \cite[Installation Instructinos]{intel:dmldoc}, meaning code developed without an alternative path will not work there. To interface with the driver and perform configuration operations, Intel's libaccel-conf \cite{intel:libaccel-config-repo} user space toolset may be used which provides a command-line interface and can read configuration files to set up the device as described previously. After successful configuration, each \gls{dsa:wq} is exposed as a character device by \texttt{mmap} of the associated portal \cite[3]{intel:analysis}. \par
|
|
|
|
Given the file permissions, it would now be possible for a process to submit work to the \gls{dsa} via either \gls{x86:movdir64b} or \gls{x86:enqcmd} instructions, providing the descriptors by manually configuring them. This, however, is quite cumbersome, which is why Intel's Data Mover Library \cite{intel:dmldoc} exists. With some limitations (like lacking support for \gls{dsa:dwq}s) this library presents a high-level interface that takes care of creation and submission of descriptors, some error handling and reporting. Thanks to the high-level-view the code may choose a different execution path at runtime which allows the memory operations to either be executed in hardware or software. The former on an accelerator or the latter using equivalent instructions provided by the library. This makes code based upon it automatically compatible with systems that do not provide hardware support. \par
|
|
|
|
\todo{finish this section}
|
|
|
|
\begin{itemize}
|
|
\item drain descriptor / drain command signals completion of preceding descriptors for fencing in non-batch submissions, in batches the ``fence flag'` can be used to ensure ordering, failures before a fence will lead to the following descriptors being aborted \cite[30]{intel:dsaspec}, \texttt{sfence} or \texttt{mfence} should be executed before pushing drain descriptor \cite[32]{intel:dsaspec}
|
|
\item cache control flag in descriptor controls whether writes are directed to cache or to memory \cite[31]{intel:dsaspec} effects on copy from DRAM > HBM unknown
|
|
\end{itemize}
|
|
|
|
\subsection{Programming Interface}
|
|
|
|
\todo{write this section}
|
|
|
|
\begin{itemize}
|
|
\item choice is intel data mover library
|
|
\item two concepts, state-based for c-api and operation-based c++
|
|
\item just explain the basics (no code) and refer to dml documentation
|
|
\end{itemize}
|
|
|
|
\section{System Setup and Configuration} \label{sec:state:setup-and-config}
|
|
|
|
\todo{write this section}
|
|
|
|
Give the reader the tools to replicate the setup.
|
|
Also explain why the BIOS-configs are required.
|
|
|
|
Setup Requirements:
|
|
\begin{itemize}
|
|
\item VT-d enabled
|
|
\item limit CPUPA to 46 Bits disabled
|
|
\item IOMMU enabled
|
|
\item kernel with iommu and idxd driver support
|
|
\item kernel option "intel\_iommu=on,sm\_on"
|
|
\item numa nodes for hbm access in bios
|
|
\end{itemize}
|
|
|
|
\cleardoublepage
|
|
|
|
%%% Local Variables:
|
|
%%% TeX-master: "diplom"
|
|
%%% End:
|