Fig. 5: Organization of the methods section with respect to the SCOPE architecture.

The first two sections reflect details about the input data. The next section gives the mathematical details for temporal self-attention, which is the workhorse for each transformer layer. Then, there are three sections on how the prediction heads, both forecasting and event prediction, are parameterized. Finally, there are two sections on the training algorithm.