Skip to main content

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

  • Article
  • Published:

Egocentric value maps of the near-body environment

Abstract

Body-part-centered response fields are pervasive in single neurons, functional magnetic resonance imaging, electroencephalography and behavior, but there is no unifying formal explanation of their origins and role. In the present study, we used reinforcement learning and artificial neural networks to demonstrate that body-part-centered fields do not simply reflect stimulus configuration, but rather action value: they naturally arise from the basic assumption that agents often experience positive or negative reward after contacting environmental objects. This perspective successfully reproduces experimental findings that are foundational in the peripersonal space literature. It also suggests that peripersonal fields provide building blocks that create a modular model of the world near the agent: an egocentric value map. This concept is strongly supported by the emergent modularity that we observed in our artificial networks. The short-term, close-range, egocentric map is analogous to the long-term, long-range, allocentric hippocampal map. This perspective fits empirical data from multiple experiments, provides testable predictions and accommodates existing explanations of peripersonal fields.

This is a preview of subscription content, access via your institution

Access options

Buy this article

USD 39.95

Prices may be subject to local taxes which are calculated during checkout

Fig. 1: Rewarded contact explains the origin and properties of body-part-centered receptive fields.
The alt text for this image may have been generated using AI.
Fig. 2: ANNs subserving interception and avoidance actions demonstrate canonical properties of peripersonal responses.
The alt text for this image may have been generated using AI.
Fig. 3: A collection of body-part-centered value fields can be used as a successor feature ψπ.
The alt text for this image may have been generated using AI.
Fig. 4: A collection of successor features can be used as an effective map of the world near the agent: an egocentric value map.
The alt text for this image may have been generated using AI.
Fig. 5: ANNs trained on multiple tasks show architectures conducive to egocentric maps.
The alt text for this image may have been generated using AI.
Fig. 6: Fitting an egocentric map to empirical data: setup and method.
The alt text for this image may have been generated using AI.
Fig. 7: Fitting egocentric maps to empirical data, part 1: overall fit and neural measures.
The alt text for this image may have been generated using AI.

Similar content being viewed by others

Data availability

All used empirical data was extracted from publicly available figures11,14,25,26,27,46,47,48. Generated data can be recreated using the code at https://github.com/rorybufacchi/EgocentricValueMaps. Source data are provided with this paper.

Code availability

All analyses were performed in MATLAB (2020a and 2022b). All code is available at https://github.com/rorybufacchi/EgocentricValueMaps.

References

  1. Moser, E. I., Kropff, E. & Moser, M.-B. Place cells, grid cells, and the brain’s spatial representation system. Annu. Rev. Neurosci. 31, 69–89 (2008).

    Article  CAS  PubMed  Google Scholar 

  2. Banino, A. et al. Vector-based navigation using grid-like representations in artificial agents. Nature 557, 429–433 (2018).

    Article  CAS  PubMed  Google Scholar 

  3. Stachenfeld, K. L., Botvinick, M. M. & Gershman, S. J. The hippocampus as a predictive map. Nat. Neurosci. 20, 1643–1653 (2017).

    Article  CAS  PubMed  Google Scholar 

  4. Russek, E. M., Momennejad, I., Botvinick, M. M., Gershman, S. J. & Daw, N. D. Predictive representations can link model-based reinforcement learning to model-free mechanisms. PLoS Comput. Biol. 13, e1005768 (2017).

    Article  PubMed  PubMed Central  Google Scholar 

  5. Behrens, T. E. J. et al. What Is a cognitive map? Organizing knowledge for flexible behavior. Neuron 100, 490–509 (2018).

  6. Bufacchi, R. J. & Iannetti, G. D. An action field theory of peripersonal space. Trends Cogn. Sci. 22, 1076–1090 (2018).

  7. Graziano, M. S. & Cooke, D. F. Parieto-frontal interactions, personal space, and defensive behavior. Neuropsychologia 44, 845–859 (2006).

    Article  PubMed  Google Scholar 

  8. Noel, J.-P., Blanke, O. & Serino, A. From multisensory integration in peripersonal space to bodily self-consciousness: from statistical regularities to statistical inference. Ann. N. Y. Acad. Sci. 1426, 146–165 (2018).

    Article  Google Scholar 

  9. Cléry, J., Guipponi, O., Wardak, C. & Ben Hamed, S. Neuronal bases of peripersonal and extrapersonal spaces, their plasticity and their dynamics: knowns and unknowns. Neuropsychologia 70, 313–326 (2015).

    Article  PubMed  Google Scholar 

  10. Makin, T. R., Holmes, N. P. & Zohary, E. Is that near my hand? Multisensory representation of peripersonal space in human intraparietal sulcus. J. Neurosci. 27, 731–740 (2007).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  11. Wamain, Y., Gabrielli, F. & Coello, Y. EEG μ rhythm in virtual reality reveals that motor coding of visual objects in peripersonal space is task dependent. Cortex 74, 20–30 (2016).

    Article  PubMed  Google Scholar 

  12. Sambo, C. F. & Forster, B. An ERP investigation on visuotactile interactions in peripersonal and extrapersonal space: evidence for the spatial rule. J. Cogn. Neurosci. 21, 1550–1559 (2009).

    Article  PubMed  Google Scholar 

  13. de Haan, A. M., Smit, M., Van der Stigchel, S. & Dijkerman, H. C. Approaching threat modulates visuotactile interactions in peripersonal space. Exp. Brain Res. 234, 1875–1884 (2016).

    Article  PubMed  PubMed Central  Google Scholar 

  14. Serino, A. et al. Body part-centered and full body-centered peripersonal space representations. Sci. Rep. 5, 18603 (2015).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  15. Hunley, S. B. & Lourenco, S. F. What is peripersonal space? An examination of unresolved empirical issues and emerging findings. Cogn. Sci. 9, e1472 (2018).

    Google Scholar 

  16. Cléry, J. et al. The prediction of impact of a looming stimulus onto the body is subserved by multisensory integration mechanisms. J. Neurosci. 37, 10656–10670 (2017).

    Article  PubMed  PubMed Central  Google Scholar 

  17. Bertoni, T., Magosso, E. & Serino, A. From statistical regularities in multisensory inputs to peripersonal space representation and body ownership: insights from a neural network model. Eur. J. Neurosci. 53, 611–636 (2021).

    Article  CAS  PubMed  Google Scholar 

  18. Roncone, A., Hoffmann, M., Pattacini, U., Fadiga, L. & Metta, G. Peripersonal space and margin of safety around the body: learning visuo-tactile associations in a humanoid robot with artificial skin. PLoS ONE 11, e0163713 (2016).

    Article  PubMed  PubMed Central  Google Scholar 

  19. Magosso, E., Zavaglia, M., Serino, A., di Pellegrino, G. & Ursino, M. Visuotactile representation of peripersonal space: a neural network study. Neural Comput. 22, 190–243 (2010).

    Article  PubMed  Google Scholar 

  20. Straka, Z., Noel, J.-P. & Hoffmann, M. A normative model of peripersonal space encoding as performing impact prediction. PLoS Comput. Biol. 18, e1010464 (2022).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  21. Bufacchi, R. J., Liang, M., Griffin, L. D. & Iannetti, G. D. A geometric model of defensive peripersonal space. J. Neurophysiol. 115, 218–225 (2016).

    Article  CAS  PubMed  Google Scholar 

  22. Pezzulo, G. & Cisek, P. Navigating the affordance landscape: feedback control as a process model of behavior and cognition. Trends Cogn. Sci. 20, 414–424 (2016).

  23. Rizzolatti, G., Fadiga, L., Fogassi, L. & Gallese, V. The space around us. Science 277, 190–191 (1997).

    Article  CAS  PubMed  Google Scholar 

  24. Fogassi, L. et al. Coding of peripersonal space in inferior premotor cortex (area F4). J. Neurophysiol. 76, 141–157 (1996).

    Article  CAS  PubMed  Google Scholar 

  25. Huijsmans, M., Haan de, M. A., Müller, B. C. N., Dijkerman, H. C. & Schie van, H. T. Knowledge of collision modulates defensive multisensory responses to looming insects in Arachnophobes. J. Exp. Psychol. Hum. Percept. Perform. https://doi.org/10.1037/xhp0000974.supp (2022).

    Article  PubMed  Google Scholar 

  26. Graziano, M. S., Hu, X. T. & Gross, C. G. Visuospatial properties of ventral premotor cortex. J. Neurophysiol. 77, 2268–2292 (1997).

    Article  CAS  PubMed  Google Scholar 

  27. Holmes, N. P. Does tool use extend peripersonal space? A review and re-analysis. Exp. Brain Res. 218, 273–282 (2012).

    Article  PubMed  Google Scholar 

  28. Iriki, A., Tanaka, M. & Iwamura, Y. Coding of modified body schema during tool use by macaque postcentral neurones. Neuroreport 7, 2325–2330 (1996).

    Article  CAS  PubMed  Google Scholar 

  29. Graziano, M. The Spaces between Us: A Story of Neuroscience, Evolution, and Human Nature (Oxford Univ. Press, 2017).

  30. Mountcastle, V. B., Lynch, J. C., Georgopoulos, A., Sakata, H. & Acuna, C. Posterior parietal association cortex of the monkey: command functions for operations within extrapersonal space. J. Neurophysiol. 38, 871–908 (1975).

    Article  CAS  PubMed  Google Scholar 

  31. Barreto, A. et al. Transfer in deep reinforcement learning using successor features and generalised policy improvement. In 35th International Conference on Machine Learning, Vol. 2, 844–853 (ICML, 2018).

  32. Dolan, R. J. & Dayan, P. Goals and habits in the brain. Neuron 80, 312–325 (2013).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  33. Cisek, P. & Kalaska, J. F. Neural mechanisms for interacting with a world full of action choices. Annu. Rev. Neurosci. 33, 269–298 (2010).

    Article  CAS  PubMed  Google Scholar 

  34. Gigliotti, M. F., Soares Coelho, P., Coutinho, J. & Coello, Y. Peripersonal space in social context is modulated by action reward, but differently in males and females. Psychol. Res. 85, 181–194 (2021).

    Article  PubMed  Google Scholar 

  35. Bertoni, T. et al. The self and the Bayesian brain: testing probabilistic models of body ownership through a self-localization task. Cortex 167, 247–272 (2023).

    Article  PubMed  Google Scholar 

  36. Friston, K., Sajid, N., Heins, C., Pavliotis, G. A. & Parr, T. The free energy principle made simpler but not too simple. Physics Rep. 1024, 1–29 (2022).

  37. Wolpert, D. M., Goodbody, S. J. & Husain, M. Maintaining internal representations: the role of the human superior parietal lobe. Nat. Neurosci. 1, 529–533 (1998).

    Article  CAS  PubMed  Google Scholar 

  38. Medendorp, W. P. & Heed, T. State estimation in posterior parietal cortex: distinct poles of environmental and bodily states. Prog. Neurobiol. https://doi.org/10.1016/j.pneurobio.2019.101691 (2019).

  39. Bufacchi, R. J., Battaglia-Mayer, A., Iannetti, G. D. & Caminiti, R. Cortico-spinal modularity in the parieto-frontal system: a new perspective on action control. Prog. Neurobiol. 231, 102537 (2023).

    Article  CAS  PubMed  Google Scholar 

  40. Arber, S. & Costa, R. M. Networking brainstem and basal ganglia circuits for movement. Nat. Rev. Neurosci. 23, 342–360 (2022).

    Article  CAS  PubMed  Google Scholar 

  41. Burgess, N. Spatial memory: how egocentric and allocentric combine. Trends Cogn. Sci. 10, 551–557 (2006).

    Article  PubMed  Google Scholar 

  42. Melzack, R. & Wall, P. D. The Challenge of Pain (Penguin, 1988).

  43. Seymour, B. et al. Temporal difference models describe higher-order learning in humans. Nature 429, 664–667 (2004).

    Article  CAS  PubMed  Google Scholar 

  44. Baliki, M. N. et al. Corticostriatal functional connectivity predicts transition to chronic back pain. Nat. Neurosci. 15, 1117–1119 (2012).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  45. Dong, W. K., Chudler, E. H., Sugiyama, K., Roberts, V. J. & Hayashi, T. Somatosensory, multisensory, and task-related neurons in cortical area 7b (PF) of unanesthetized monkeys. J. Neurophysiol. 72, 542–564 (1994).

  46. Ronga, I. et al. Seeming confines: electrophysiological evidence of peripersonal space remapping following tool-use in humans. Cortex 144, 133–150 (2021).

    Article  PubMed  Google Scholar 

  47. Quinlan, D. J. & Culham, J. C. fMRI reveals a preference for near viewing in the human parieto-occipital cortex. Neuroimage 36, 167–187 (2007).

    Article  CAS  PubMed  Google Scholar 

  48. Holt, D. J. et al. Neural correlates of personal space intrusion. J. Neurosci. 34, 4123–4134 (2014).

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  49. Ferri, F., Tajadura-Jiménez, A., Väljamäe, A., Vastano, R. & Costantini, M. Emotion-inducing approaching sounds shape the boundaries of multisensory peripersonal space. Neuropsychologia 70, 468–475 (2015).

    Article  PubMed  Google Scholar 

  50. Taffou, M. & Viaud-Delmon, I. Cynophobic fear adaptively extends peri-personal space. Front. Psychiatry 5, 3–9 (2014).

    Article  Google Scholar 

Download references

Acknowledgements

We thank L. Bonini, N. Burgess, F. Cacucci, P. Neri and G. Vallortigara for their comments on earlier versions of the manuscript, and K. Shao and S. Perovic for discussions and contributions to the figures. The present study was supported by the European Research Council (ERC Consolidator Grant PAINSTRAT to G.D.I.), a fellowship from the Shanghai Municipal Human Resources and Social Security Bureau (no. E35CN31A21 to R.J.B.), a fellowship from the Italian Academy for Advanced Studies of Columbia University (to G.D.I.) and the Shanghai Municipal Science and Technology Major Project (grant no. 2019SHZDZX02 to N.L).

Author information

Authors and Affiliations

Authors

Contributions

R.J.B. and G.D.I. conceived the project. R.J.B. did the modeling and analysis. R.J.B. and G.D.I. created the figures. R.J.B., R.S., A.M.F., R.C. and G.D.I. planned the analysis. R.J.B. and G.D.I. wrote the original draft. R.J.B., R.S., A.M.F., Y.M., N.L., R.C. and G.D.I. reviewed and edited the MS. Y.M., N.L. and G.D.I. acquired the funds.

Corresponding authors

Correspondence to Rory John Bufacchi or Gian Domenico Iannetti.

Ethics declarations

Competing interests

The authors declare no competing interests.

Peer review

Peer review information

Nature Neuroscience thanks Andrea Serino and the other, anonymous, reviewer(s) for their contribution to the peer review of this work.

Additional information

Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Extended data

Extended Data Fig. 1 Bodypart-centred value fields arise even if touch is not directly rewarded.

a) In the main text we assumed that touch resulted in immediate reward, for the sake of simplicity. However, bodypart-centred fields can also arise if touch itself is not directly rewarded, but is a prerequisite to obtain reward at a later stage. To demonstrate this, consider the situation in which an agent must first catch an object (here, an apple) with its limb (step 1) before bringing it to its mouth (step 2) and finally obtaining the reward by ‘eating’ it (step 3). b) Value fields as a function of object and mouth position. Each column shows Q-values for a specific mouth position. Right-most column shows the average Q-values across all mouth positions. Line plots to the bottom and left of each heatmap are average Q-values along the y and x axes, respectively. In this environment, value fields still emerge around the limb, because contact is a step along the path to reward. The field shape and magnitude also depend on the position of the mouth, because intercepting the ‘apple’ when the mouth is near the hand leads to more immediate reward, and is therefore more valuable.

Extended Data Fig. 2 Individual neural response fields in all architectures used for ANN model 1.

a) Each colour map shows the activity of a single network unit as a function of the stimulus spatial position relative to the body part (body-part location is indicated as white circles). The three main rows indicate different network architectures. Full-colour plots show units classified as bodypart-centred; greyed-out plots show units not classified as bodypart-centred. Within each ANN, the proportion of bodypart-centred units per layer increased as a function of layer depth. b) Same as (a), but for ANNs that controlled two body parts instead of 1.

Extended Data Fig. 3 Modularity emerges only when multiple tasks are learned successfully, regardless of artificial neuron type.

a) An ANN trained on simultaneous interception and avoidance tasks naturally adopts a modular, task-specialised structure. This example ANN consists of fewer neurons than the modular ANNs reported in the main text (see Fig. 5 and Supplementary Methods 3.2.6). Individual units are classified as threat- or goal-preferring (red and blue, respectively). b) Structural modularity in ANNs, split by the neuron types that compose the ANNs (see Supplementary Methods 3.2.6 and Supplementary Results 4.1.2). Rows indicate neuron type, while columns indicate regularization method. The histograms above each line plot show the amount of sub-network structure (indexed by a t-statistic; see Supplementary Methods 3.2.6). Histograms to the left of each line plot show network performance (indexed by expected reward per unit time). Light blue shaded area indicates above-chance sub-network structure. Network types with above-chance performance also had above-chance sub-network structure, regardless of regularization method. c) Across neuron types and regularization methods, the amount of sub-network structure (x-axis) predicted the performance of the network (y-axis). Colours indicate neuron types. Dashed line indicates zero. Light blue shaded area indicates above-chance sub-network structure. Rho and p-value result from a Pearson's correlation test. d) The same ANN as in (a) also naturally encodes task-specific variables in orthogonal spaces. Scatter plots show activity from the first 3 principal components. Information related to goal-proximity (black-to-blue scatter plot) is orthogonal to information about threat-proximity (black-to-red scatter plot). e) Orthogonality of goal and threat proximity coding in the same networks described in (b). Coloured histograms show the angle between goal- and threat-coding in the 1st 3 PCs. Light blue areas indicate angles within 15° of 90°. Grey histograms indicate null distributions (1000 permutations). f) The degree of task-encoding orthogonality in the first 3 principal components (x-axis) predicted the performance of the network (y-axis). Dashed line indicates the absolute dot product equivalent to a 75° angle difference. Light blue shaded area indicates angles within 15° of 90°. Rho and p-value result from a Pearson correlation test. g) Across neuron types and regularization methods, the amount of sub-network structure (x-axis) is predictive of the ability of the network to reconstruct Q-values for novel tasks (y-axis; correlation coefficient between original and reconstructed Q-values; see Supplementary Methods 3.2.6). The novel tasks were the same as those described in Fig. 4a of the main text. Rho and p-value result from a Partial correlation test, which factors out the effects of performance (reward per unit time) and task-encoding orthogonality. h) Across neuron types and regularization methods, the amount of task-encoding orthogonality (x-axis) is also predictive of the ability of the network to reconstruct Q-values for novel tasks (y-axis). Colours indicate neuron types. Rho and p-value result from a Partial correlation test, which factors out the effects of performance (reward per unit time) and network structure.

Extended Data Fig. 4 Fitting egocentric maps to empirical data, part 2: setup and behavioural measures.

a) To create a model egocentric map around the upper body, we created a default set of voxels on the surface of the hand, head, and trunk, which offered reward upon contact (coloured cubes). Experiment-dependent variations of these rewarded voxels are shown in b-e. For extended data fitting methodology, see Methods 'Empirical data fitting' and Supplementary Methods 3.4.1. b) To model arm-centred peripersonal neurons in macaques (Fig. 7a), we created two sets of voxels, each simulating one of the two arm positions during the experiment. c) In a subset of human experiments (I,j,l,m,n,o,p; Fig. 7e–h), the hand was held in front of the chest, instead of to the side. d) To model experiments with variable stimulus valence (n,o,p), we additionally optimised the negative reward offered by contact (Methods, 'Empirical data fitting' & Supplementary Methods 3.4.1). The optimised reward values implied by the stimuli of differing valence are shown in q. e) To model tool use (m; Fig.7h), we also rewarded contact with the tip of a tool. The two tool-related experiments used a stick and a rake, which we respectively modelled with 1 (green) or 5 (yellow) voxels. f-p) Empirical data (right in each panel) and model fits (left in each panel) from experiments in which behavioural peripersonal measures were collected. For data and fits of neural measures, see Fig. 7c–h. Colours indicate experimental conditions. Shaded surfaces on figurines show body parts included in the egocentric map. Error bars show SEM. For a detailed description of each fitted experiment, see Supplementary Methods 3.4.1. f) See Supplementary Methods 3.4.1.6 for details. g) See Supplementary Methods 3.4.1.6 for details. h) See Supplementary Methods 3.4.1.6 for details. i) See Supplementary Methods 3.4.1.6 for details. j) See Supplementary Methods 3.4.1.6 for details. k) See Supplementary Methods 3.4.1.6 for details. l) See Supplementary Methods 3.4.1.6 for details. m) See Supplementary Methods 3.4.1.7 for details. n) See Supplementary Methods 3.4.1.8 for details. o) See Supplementary Methods 3.4.1.9 for details. p) See Supplementary Methods 3.4.1.10 for details. q) Best-fitting negative reward magnitudes for the stimuli of differing valence from experiments displayed in n,o, and p.

Source data

Extended Data Fig. 5 Comparison to alternative models.

a) We fitted three model families to an empirical dataset combined from 23 published experiments across 10 different research groups. The ‘Egocentric maps’ family (top three models) is the main topic of this paper, and the ‘Q-fields (Q-learning)’ model is the specific model described in the main text. The ‘Monotonous decay’ family (middle three models) contains purely empirical models that attempt to describe the data, but without having a theoretical a-priori reason for being appropriate models. The ‘Perceptual models’ (bottom two models) have previously been used to fit individual datasets, and are largely based on the notion that peripersonal fields arise due to uncertainty in visual and auditory input, while estimating the probability that the source of the visual input makes contact with the body. We calculated all quantities for each 5 × 5 × 5cm voxel around the upper body, and fit them to the data with at least the same number of parameters as we used for Q-value fitting. The exponential and linear falloffs required two additional parameters, to fit the size and slope of the receptive fields. We parametrised the uncertainty necessary for the perceptual models by taking the same values as reported in17,20. b) Mathematical description of each model. For the ‘Egocentric maps’ models family, we display the update equation for the Q values, and underline the part of the equation that is unique to each of the three models. c) Summed error when each model is fitted to the empirical data (red line). The error of all models other than egocentric maps is larger than the error expected from a model that appropriately describes the generative mechanism behind the data (blue distribution). Models with a summed error corresponding to p < 0.05 (using the variant of chi-squared goodness of fit testing described in Methods 'Empirical data fitting') can be confidently rejected as explanations of the data. d) Metrics of fit quality relative to the ‘Q-fields (Q-learning)’ model. The normalised error (left y axis, red) is the summed error from (c) scaled by the variability of the data. Purple bars show the difference between the AIC and BIC (Akaike and Bayes Information Criterion, respectively) of each model vs the ‘Q-fields (Q-learning)’ model (right y axis, purple). A difference of >10 for AIC and BIC (indicated by dashed black lines) is commonly taken to indicate that the considered model can be rejected in favour of the reference model.

Supplementary information

Supplementary Information (download PDF )

Supplementary Methods, Results, Discussion, Figs. 1–5, Videos and References.

Reporting Summary (download PDF )

Source data

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Bufacchi, R.J., Somervail, R., Fitzpatrick, A.M. et al. Egocentric value maps of the near-body environment. Nat Neurosci 28, 1336–1347 (2025). https://doi.org/10.1038/s41593-025-01958-7

Download citation

  • Received:

  • Accepted:

  • Published:

  • Version of record:

  • Issue date:

  • DOI: https://doi.org/10.1038/s41593-025-01958-7

Search

Quick links

Nature Briefing

Sign up for the Nature Briefing newsletter — what matters in science, free to your inbox daily.

Get the most important science stories of the day, free in your inbox. Sign up for Nature Briefing