Cloud native EDA tools & pre-optimized hardware platforms
Compute Express Link (CXL) is the latest specification in interconnect technology for high bandwidth devices. It provides high-speed, efficient connectivity from CPUs to other components of the high-performance computing platform.
CXL’s coherent memory access capability between a host CPU and a device, such as hardware accelerators, meets the requirements for handling data and compute intensive workloads in next-generation designs by leveraging the advanced capabilities of PCIe architecture.
Let’s explore the types of CXL devices and the verification challenges that are unique to CXL, like maintaining the cache coherency between a host CPU and an accelerator. For an initial overview of this specification, check out our previous blog here.
Type 1 CXL Device
Type 2 CXL Device
Type 3 CXL Device
A CXL.cache/mem design has to maintain cache coherency between the host and device agent caches and their corresponding memory. The Bias Based Coherency model for a Type 1 CXL device defines two states for device attached memory – the host bias and the device bias. Each possess their own verification challenges.
CXL acts as a high-performing I/O interconnect system, trying to ensure reliable and efficient access to memory distributed throughout various components. Some of these components optimize performance by making use of the local cache and reduce the overhead of memory access. To support this type of configuration, CXL.cache protocol ensures that data held across the components in either memory or local cache remain coherent and consistent to each component. Device components in CXL are typically used as accelerators for computationally intensive applications, and hence contain a local cache. So, if the host component wants to access the same location of memory, it can then force the device to evict the line from its local cache and update the corresponding memory location depending on the cache line state.
The CXL.cache protocol defines interaction between the device and host as a number of requests that each have at least one associated response message and sometimes a data transfer. The interface consists of three channels in each direction: Request, Response, and Data. The channels are named for their direction – D2H (Device to Host) and H2D (Host to Device).
A CXL based system uses the Bias Based Coherency Model to improve memory access performance. When the device-attached memory is in a host bias state, the device accesses it like a regular host-attached memory. If the device needs to access the device-attached memory it first sends a request to the host which will resolve coherency for the requested line.
Alternatively, when the device-attached memory is in device bias state, the device is guaranteed that the host does not have the line cached. In this case, the device can access it without sending any transaction to the host.; however, the host sees a uniform view of device-attached memory regardless of the bias state.
In both the host bias state and the device bias state, coherency is always preserved for device-attached memory.
CXL.cache H2D Snp Transaction exchange:
A Type 3 device is primarily a memory expander for the host memory. The device operates over CXL.mem to service requests sent from the host. Two types of flows are supported for Type 3 devices – Read and Write flows.
Synopsys is the market leader for Design IP and Verification IP for CXL, contributing significantly to the evolution of the CXL ecosystem. Stay tuned for more deep dives on this new CXL specification, including topics like additional transaction types, layered architecture and the verification requirement/challenges of CXL designs.
For more information, please visit http://synopsys.com/vip