Multicell Capacity
CPU core usage for multicell benchmark:
On Grace Hopper:
1 isolated physical Grace core for core-locked PTP applications (phc2sys+ptp4l)
10 additional isolated Grace cores for the other core-locked cuphycontroller threads
On x86 based targets:
Without hyperthreading using “1+6+fractional” x86 cores. The shorthand “1+6+fractional” x86 cores is defined as follows:
1 isolated physical x86 core for core-locked PTP applications (phc2sys+ptp4l) and the core-locked cuphycontroller L2A H2D prepone thread
6 aditional isolated physical x86 cores for the other core-locked cuphycontroller threads
A fraction of a shared floating x86 core for non-core-locked cuphycontroller threads
Additionally, the tested L2 timeline is as follows:
FAPI SLOT.indication for Slot N is sent from L1 to L2 at the wall-clock time for Slot N-3 (i.e. 3 slot advance).
For the DDDSUUDDDD TDD pattern with 0-based slot numbering, L2 has up-to-500us, for slot%10 in {2,3,4,5,6}, from SLOT.indication to deliver all FAPI PDUs for Slot N.
For the DDDSUUDDDD TDD pattern with 0-based slot numbering, L2 has up-to-250us, for slot%10 in {0,1,7,8,9}, from SLOT.indication to deliver all FAPI PDUs for Slot N.
As of 24-1:
Supports 500us L2 processing budget and 7 beam pattern as defined below using 100MHz patterns 59c and 60c:
On Grace Hopper:
BFP9: 20 4T4R avg cells
On discrete A100/CX6 cards:
BFP9: 4 4T4R peak / 8 4T4R avg cells
On A100X:
BFP9: 5 4T4R peak cells / 10 4T4R avg cells
BFP14: 4 4T4R peak cells / 8 4T4R avg cells
On AX800:
BFP9: 8 4T4R peak cells / 16 4T4R avg cells
BFP14: 8 4T4R peak cells / 16 4T4R avg cells using both FH ports of the AX800
while respecting the following configuration for 7 beam traffic patterns:
7-beam config |
Configuration (4 UL streams RU->DU) |
|
---|---|---|
Peak |
Average |
|
Compression | BFP9 and BFP14 | BFP9 and BFP14 |
Max PxSCH PRB | 270 | 132 |
DL Throughput/cell | 1469.14 Mbps | 523.10 Mbps |
UL Throughput/cell | 196.70 Mbps | 79.91 Mbps |
Peak DL Fronthaul Bandwidth / cell | 11.06 Gbps BFP14 | 5.46 Gbps BFP14 |
7.14 Gbps BFP9 | 3.58 Gbps BFP9 | |
Peak UL Fronthaul Bandwidth / cell | 11.88 Gbps BFP14 | 6.34 Gbps BFP14 |
8.03 Gbps BFP9 | 4.57 Gbps BFP9 | |
SSB slots | Frame 0 & 2: 0,1,2,3 | Frame 0 & 2: 0,1,2,3 |
#SSB per slot | Frame 0 & 2: 2,2,2,1 | Frame 0 & 2: 2,2,2,1 |
TRS slots | Frame 0-3: 6,7,8,9,10,11 | Frame 0-3: 6,7,8,9,10,11 |
Frame 0 & 2: 16,17 | Frame 0 & 2: 16,17 | |
TRS Symbols | Even cells: 6,10 | Even cells: 6,10 |
Odd cells: 5,9 | Odd cells: 5,9 | |
CSI-RS slots | Frame 0: 8,10,16 | Frame 0: 8,10,16 |
Frame 1: 6,8,10 | Frame 1: 6,8,10 | |
Frame 2: 6 | Frame 2: 6 | |
CSI-RS Symbols | Even cells: 12 | Even cells: 12 |
Odd cells: 13 | Odd cells: 13 | |
PDSCH #DCI | 12 (6 DL + 6 UL per slot) | 12 (6 DL + 6 UL per slot) |
UE/TTI/Cell | 6 per DL slot, 6 per UL slot | 6 per DL slot, 6 per UL slot |
UCI on PUSCH HARQ+CSIP1+CSIP2 (bits) | 4+37+5 | 4+37+5 |
PUCCH format | 1 | 1 |
PUCCH payload (bits) | 18 | 18 |
PRACH format | B4 | B4 |
PRACH slots | Frame 0-3: 5, 15 | Frame 0-3: 5, 15 |
PRACH occasions | Slot 5: 4, Slot 15: 3 | Slot 5: 4, Slot 15: 3 |
Notes:
Stated performance achievement and CPU core count usage is for L1 workload only (additional non-L1 workloads in E2E setting may have an impact on the achieved performance and/or CPU core count usage)
Performance achievement is measured by running L1 in steady-state traffic mode (e.g. impact of workloads such as cell reconfiguration on other cells is not captured)