Introduction

The Late Iron Age, or La Tène period (500–0 BCE), in Central Europe was marked by major societal transformations, culminating towards an end of the third century BCE1. In this period, the evidence points to the rise of large, nucleated settlements focused on craft specialisation, intensification. Integration of agglomerations into expansive supra-regional trade and communication networks brought in innovations in traditional technologies such as metalworking2, new materials—glass3, leaded bronze, and silver2,4, along with new decorative styles5 and ritual practices1. The introduction of coinage6 shaped the nature of trade relations, especially with the Mediterranean, but also within wider Transalpine area. By the late third century BCE, socio-economic structures had reached a level of complexity broadly comparable to pre-industrial or traditional modern societies. An integral part of the transformation was also a development of agricultural strategies. The traditionally cultivated crop spectrum was quite diverse, and varied regionally, with cereals typically dominated by barley and wheats 7,8,9,10,11. Archaeobotanical and osteological analyses from agglomerations noticed shifts toward more productive systems capable of supporting denser populations and growing trade10,12. A notable trend was the increasing reliance on millet as one of the staple crops, though its occurrence varied regionally as well as chronologically10,13,14,15.

Understanding how past communities adapted their food systems to environmental and societal challenges is essential for informing modern sustainability strategies. Millet, in particular, and its varying use across different periods and environments has proven valuable for examining broader societal dynamics, such as migration, production intensification, and social differentiation, that underlie agricultural change16,17,18,19,20. However, interpreting these variations remains challenging without sufficient contextual and supporting evidence. Traditionally, past crop use is studied through archaeobotanical macro-remains, which provide insight into dietary patterns and plant food choices shaped by environmental or cultural factors21. Stable isotope analysis, particularly δ13C in human collagen, offers an independent proxy, reflecting the balance of C₃ and C₄ plants (or marine inputs) in past human and animal diets22,23. C₃ crops (e.g. wheat, barley, legumes), dominating mesic woodland biomes4,8, typically show δ13C values from –37‰ to –20‰, influenced by factors like aridity or the canopy effect24,25. C₄ plants, on the other hand, such as millets, are typical for tropical grassland or savannah biomes19,26 and range from –15‰ to –10‰ in modern crops27, and around –12.7‰ to –11.4‰ in fossil specimens from Europe13,19. C₄ plants are more tolerant of arid conditions, displaying greater resilience and adaptability to water availability, and have shorter growing seasons, making them resilient in marginal environments18,28,29. In pre-industrial mid-latitude Europe, millets, specifically Panicum miliaceum and Setaria italica, were the only C₄ crops widely cultivated30. Their distinctive isotopic ‘fingerprint´ allows for reliable identification in diet. For C₄ signals to appear in collagen, millet-derived protein must make up at least 20% of the diet31. Given a trophic enrichment factor for carbon of up to 1‰32, a δ13C collagen threshold of − 18‰ (VPDB) is conventionally used to indicate substantial millet intake 13,18,19,33,34,35,36. Values near this mark suggest a mixed C₃/C₄ diet, while higher values reflect C₄ dominance18,37,38.

Recent scholarship has explored the cultural and economic roles of millet across Eurasia18,20,39,40, where its spread during the Late Bronze Age (ca. 1600–1200 BCE)20,36,38,41,42,43,44,45,46,47 has been framed as an episode of ‘food globalisation’—a reflection of intensified cultural connectivity and innovation18,40,43,48,49,50. Consequently, millet emerged as a symbol of innovation and adaptability, with its cultivation considered a critical driver of societal transformations and agricultural advances that enhanced subsistence stability in prehistoric communities18,19. It became a dietary mainstay across many regions, as shown by both archaeobotanical and isotopic evidence7,18,36,41,51,52,53,54, though its prominence declined in some areas after the Bronze Age7,36,38, arguably due to climate change16. In later periods, millet periodically resurfaced in specific socio-cultural contexts13,34,35,41,55,56. Besides direct link to climate fluctuations in certain periods16,57, integration of millet into agricultural strategies appears to reflect distinct socio-economic settings or culturally mediated preferences. For example, millet held a central role in Early Medieval agricultural systems of Southern58,59,60,61 and Eastern Europe46,62,63, while remaining largely absent among Germanic communities in Central and Western Europe50,64,65,66.

Unlike the Bronze Age, when millet first appeared as a novel crop, the La Tène period allows for the examination of shifts within an already established crop spectrum. This context enables a more detailed assessment of how environmental and socio-economic factors influenced food production, in ways that are arguably closer to more recent historical patterns. Regional and temporal fluctuations in millet use underscore its value as a proxy for agricultural resilience and socio-economic adaptation. When viewed across long timespans, isotopic shifts in human diets can be aligned with archaeobotanical trends to reconstruct nuanced cultivation and consumption strategies. This dual approach, when applied at the highest achievable temporal, spatial, and social resolution, offers key insights into how environmental, technological, and cultural forces shaped past subsistence systems, with meaningful implications for food security in today’s globalised yet ecologically diverse world.

In this study, the significance of millet within late Iron Age agriculture is being evaluated using a multi-proxy approach of combined analysis of archaeobotanical and isotopic evidence. This multi-proxy approach is applied here for the first time, made possible by the availability of sufficiently robust and analytically comparable datasets from three regions in Central Europe —Bohemia, Moravia, and Slovakia (Fig. 1). Settlement-based archaeobotanical data were classified into distinct chronological phases to align with stable isotope evidence from human and animal remains (see Methods). Previous research identified regional variation in crop use, with barley, spelt wheat, free-threshing wheat, einkorn, and millet forming the primary agricultural spectrum8, with proportions influenced by geography and altitude9,10,14,15,51. Isotopic data, available for ca. 400–200 BCE, suggest a general reliance on C₃ crops33. However, several sites show a marked shift towards higher δ13C values during the third century BCE56, indicating increased millet consumption. Comparable trends have been observed in Switzerland13,67 and Italy37, but broader synthetic evaluations remain lacking.

Fig. 1
figure 1

© Esri and NASA, NGA, USGS, GUGiK, ŠOP SR, Esri TomTom, Garmin, FAO, METI/NASA; graphic by A.D.).

Map of the area under study (“BOH”–Bohemia, “MOR”–Moravia, “SLO”–Slovakia) with Iron Age sites with botanical (A) and stable isotope (B) data. The lists of archaeological contexts (sites) and associated datasets are provided in Tables S4 and S9, and Fig. S3 and S4. (source maps: the Earth created with MapChart.net, CC BY-SA 4.0; https://creativecommons.org/licenses/by-sa/4.0/; basemaps in A and B insets: World Topographic Map by

To address this, we analyse new and previously published stable isotope data from burials contextualised by grave goods and demographic information. Statistical evaluation of both archaeobotanical and isotopic datasets, at the highest resolution permitted by the archaeological record, seeks to identify correlating temporal, geographical, and social patterns. Specifically, we test whether the isotopic shift in the third century BCE corresponds with a systemic increase in millet cultivation across the study region. The results are discussed in relation to changing food production strategies and the socio-economic organisation of La Tène societies undergoing societal transformation. By comparing production trends with local constraints, we explore how rural communities negotiated supra-regional processes while adapting to specific ecological and social settings. Finally, we address methodological considerations in integrating archaeobotanical data with dietary patterns inferred from stable isotope analysis.

Results

Archaeobotanical analysis

The data clearly demonstrate that millet is a relatively constant crop throughout the analysed period (Table S1, Fig. S5, S9). Its occurrence and abundance slightly fluctuate in time, and it is more common at sites in the eastern part of the study area.

The Detrended Correspondence Analysis (DCA) performed on various cereal data matrices (original, transformed, and reduced) did not reveal any definitive temporal or regional trends. In the DCA plot based on percentage values (Fig. 2), millet-rich samples were associated with low positive values on the horizontal axis (left side of the scatter), whereas free-threshing wheat-rich samples clustered at the opposite positive end. Coding samples by region and dating did not demonstrate clear separations or trends, although millet-rich samples were primarily associated with Slovakia and, to a certain extent, also Moravia.

Fig. 2
figure 2

DCA plot of cereal crop assemblages dated to the La Tène period, showing cereal proportions (a), determined cereal finds (b), regional distribution (c), and chronological phase (d), (Graphic by M.H.).

The statistical testing using parametric one-way ANOVA was conducted using two matrices; one based on ubiquity, representing the frequency of millet occurrence across sites, and the other on percentage, calculated as the minimum number of individual (MNI) millet remains per site in relation to the overall crop spectrum. In each analysis the results are based on testing four matrices in total: two based on the original data and two on the modelled data. All the results and p values are listed in Table 1.

Table 1 Results of a series of one-way ANOVA tests with post-hoc Tukey’s Honest Significant Difference and Bonferroni-corrected tests run on various matrices of Panicum data.

The results on ‘ubiquity’ data, where data reduction was based on dating accuracy of sites, revealed statistically significant chronological differences either between the early (500–330 BCE) and the middle (330–180 BCE) phase or between the middle or the late (180 – 0 BCE/CE) phase, depending on which matrix was tested. Testing for geographical variability revealed statistically significant differences in all matrices used. Post hoc tests revealed that the variation lay between Bohemia and Slovakia, while in the case of modelled data, differences were also observed between Bohemia and Moravia. In summary, the analysis of ubiquity data indicates that geographical differences have a stronger influence on the patterns observed than chronological variation. Panicum is more frequently found in samples from Slovakia, and to a certain extent also from Moravia, compared to Bohemia. However, from the statistical point of view, the frequency or abundance of millet does not significantly change over established chronological phases within the individual regions. The absence of chronological variation when more sites are included suggests that unrecognised biases, such as uneven sample distribution, may be affecting the results.

The results of the one-way ANOVA on ‘percentages’ data, where data reduction was based both on MNI values (> 30 and 50 determined cereal finds, respectively) and dating accuracy, revealed no significant chronological differences suggesting that the proportion of millet relative to other cereal crops does not change significantly over time when the whole regions are analysed. Statistically significant differences between the regions were observed only when unreduced data were tested, however, these results are likely biased due to the inclusion of samples with low find counts. Consequently, it is likely that the percentage or abundance of Panicum does not differ significantly across regions when using the reduced data matrices. In general, while geographical patterns emerge more clearly in ubiquity data, the percentage-based analysis suggests that regional differences in millet abundance are not statistically significant. For percentage data, the lack of clear trends may stem from the relative nature of percentages, which are influenced by sample size. Percentages do not differentiate between actual changes in millet abundance and apparent changes caused by varying proportions of other cereals.

In conclusion, the statistical analysis of archaeobotanical data indicates minimal or no statistically significant difference in the ubiquity—frequency of occurrence—of millet between the individual chronological phases. On the other hand, a geographical gradient is evident, with millet appearing more frequently in Slovakia in the East than in Bohemia in the West, while Moravia occupies an intermediate position. No statistically significant differences were observed in the proportions of millet within site assemblages, as millet-rich and millet-poor samples were present across all chronological phases and throughout the entire geographical gradient. Future studies should address potential biases and use more refined metrics, such as seed density per litre of sediment, to improve data reliability and resolution. Additionally, site-specific or more detailed sub-regional resolution is essential, as demonstrated by spatial plots for each site (Fig. S5S12). At this refined scale, geographical and chronological variation becomes evident, revealing differences on a local scale. The ubiquity data indicate a peak in millet presence during the middle period, with percentage values highlighting this intensity particularly in central Bohemia and Moravia, and its continuation into the late period in Slovakia.

Results from the Representativeness index (RI) analysis (Table S2) align with findings derived from ubiquity analysis. When comparing aggregated data by individual territories, millet shows the highest RI value in Bohemia, with Moravia and Slovakia both at comparable, much lower levels. When examining millet in relation to other crops, it has a relatively smaller role in Bohemia compared to the two eastern regions, where the RI for millet is similar to barley and spelt. This apparent discrepancy largely arises from the greater abundance of archaeobotanical data from Bohemia (Fig. 3). Thus, in this particular context, the RI primarily reflects data robustness. Aggregated data categorised by chronology indicate that the highest RI values for millet occur during the middle period. This is chiefly because the middle period dataset is statistically the most robust, with relatively balanced contributions from each of the three studied regions.

Fig. 3
figure 3

Results of the Representativeness index categorised by chronological period and geographical region and calculated for five principal staple cereals in the analysed territory. Differences in the abundance of archaeobotanical data are particularly evident for the early period between Bohemia compared to Moravia with Slovakia, (graphic by A.D., A.P.).

A more detailed chronological and geographical comparison, evaluating the position of millet among the five principal staple cereals (Fig. 3), reveals a comparatively diminished role for millet in the early period crop spectrum in Bohemia. In contrast, millet appears relatively more prominent in Moravia and Slovakia during the same phase, although the small sample sizes limit the reliability of this observation. For the middle period, millet emerges as the second most important cereal after barley in Moravia. In Bohemia, it ranks behind barley and the wheat species (except spelt), whereas in Slovakia millet and barley both follow spelt in importance. During the late period, Slovakia provides the most abundant data, indicating millet’s importance to be broadly comparable with other staple cereals. Conversely, in Bohemia and Moravia, the available evidence suggests a notably diminished role for millet, placing it behind all other staple crops, although this conclusion remains tentative due to limited data.

Isotope analysis

For the entire human dataset (Fig. 4, 5; Table S9, descriptive statistics in Table S11, for a detailed report on sample categorisation and data analysis, see Supplementary material “SI_analysis”), the difference in average δ13C values between the beginning and end of the investigated period (400–180 BCE) is Δ13C = 0.61‰. This difference becomes more pronounced when comparing the first and third quartiles (Δ13C q1 = 0.622‰, Δ13C q3 = 0.652‰). These observations are statistically significant (Mann–Whitney U test, p = 0.000001).

Fig. 4
figure 4

Chronological differences in δ13C values between the 4th and the third century BCE for the human dataset represented by KDE plots: per century (a) per relative chronological phase (b). (graphic by S.K., A.D.).

Fig. 5
figure 5

Differences in the shift towards higher mean δ13C values between the earliest and latest chronological phases (LT B1 vs. LT C1) for individual cemeteries. The colour legend indicates median Δ13Cearly-late values (positive values mean positive shift, while negative values mean the negative shift), while circles and labels denote site sizes and the number of data per site. Double circles mark sites with n > 30 graves (graphic by A.D.).

The dataset structure reveals that the early period (400 – 300 BCE) is distinctively variable, including numerous outliers extending to lower δ13C values especially in Moravia (Fig. S1, S1.2), where the lowest outliers observed are < -21‰. On the contrary, a group of higher δ15N values and a few distinctively positive carbon outliers (> -16‰) are observed in Bohemia (Fig. S1, S1.1). Bohemian sites exhibit greater number of outliers (usually from two to six per site), while Moravian sites have wider ranges of values. This heterogeneity indicates diverse dietary practices and possibly also increased territorial mobility typical for this period68. The δ13C values become increasingly homogeneous (with non-outlier range about 2.5‰, from − 19.5 to − 17.2‰; Fig. S13, S14, Tables S13, S14) towards the later (´middle´—300 – 180 BCE) period (Mann–Whitney U test, p =  < 0.00005), suggesting perhaps a more structured approach to plant consumption. Later period outliers are almost exclusively in higher δ13C direction (with values > -17‰). Especially in the territory of Moravia, the gradual positive shift is more pronounced than in Bohemia (no data are available for Slovakia), (Fig. S14, S15, Tables S15, S16). A statistically significant difference is observed between the latest (LT C1) phase and all the older phases (Kruskal–Wallis test: LT B1–C1, p = 0.000001; LT B2a–C1, p = 0.00017; LT B2b–C1, p = 0.007, Table S14), though the change is visually observable throughout the third century BCE (Fig. S13C). However, this trend is not universally observed across regions (Figs. 5, S1, S14, S15, Tables S16, S17) and varies significantly between individual sites (Fig. S1.1S1.3).

Geographically, the largest chronological shift towards higher δ13C values is particularly evident in the three largest cemeteries of Bohemia and Moravia (Fig. 5, S1.1.2, S1.1.5, S1.2.4). Results from other sites show considerable variability, although a general trend towards greater shifts is apparent in Moravia (Fig. 5). There, the patterns of change are more consistent compared to Bohemia, where some sites even exhibit negative shifts. On average, the positive shift in δ13C values at sites where a change is recorded ranges between 0.5‰ and 1.9‰. However, even at sites showing the strongest positive shifts, maximum values do not systematically exceed − 18‰ (VPDB), the threshold generally indicating more significant dietary reliance on millet. The concentration of positive δ13C shifts at large cemeteries may equally be influenced by sampling bias, as smaller sites provide fewer data points, however, it is important to note that even among large cemeteries (n > 30), δ13C values remain variable, with only three sites exhibiting a clear positive shift of around 1‰ (Figs. 5, S1.1.2, S1.1.5, S1.2.4).

Statistical analysis of demographic age and biological sex groups (Fig. S16S19, Table S18S22) indicated that these factors generally had no significant influence on δ13C values. The sole exception was a statistically significant difference between biological males and females in the third century BCE (Mann–Whitney U test, p = 0.0076), with females exhibiting slightly higher δ13C values, suggesting greater consumption of C₄ plant protein (mean = –18.5‰ for females and –18.9‰ for males). However, this pattern was not consistent across relative chronological phases, and the high proportion of individuals with undetermined sex introduces potential bias into the analysis.

The subsequent analysis focused on potential dietary differences between social groups, as defined by the categorisation of grave good assemblages (SI_analysis). The results (Fig. S20S23, Table S23S25) indicate that the chronological shift in δ13C values is particularly pronounced in Group 2 (the “common group”). A statistically significant increase in δ13C values was observed for this group (Mann–Whitney U test, p = 0.004), with values regularly exceeding –17‰ from the LT B2b–C1 phase onwards (mid-third century BCE). Although Group 2 retains the widest overall range of δ13C values, the upward shift suggests increased consumption of C₄ plants during this period. In contrast, Groups 1A and 1B display more conservative dietary patterns, with δ13C values consistently remaining below the –18‰ threshold, indicating a continued reliance on C₃ plant protein. While a slight increase in δ13C values is visible in graphical outputs (Fig. S23), the differences are not statistically significant (Mann–Whitney U test: Group 1A, p = 0.102; Group 1B, p = 0.1). Social group differences become statistically significant in the LT C1 phase for Group 2 (Kruskal–Wallis test, p = 0.01), although the most visually apparent divergence begins in the preceding LT B2b–C1 phase (Fig. S22, S23, Table S24). It should be noted, however, that a substantial number of undated samples in Group 2—due to the absence of datable grave goods—may introduce bias into the analysis.

Finally, available animal isotopic dataset (Table S10) revealed that animal δ13C values typically peaked below -19‰, often reaching − 21‰, with the lowest recorded value at -22.2‰. This pattern indicates that animals were predominantly fed a C₃-based diet. Considering the trophic shift, differences of around 2‰ between the mean human and animal δ13C values across individual sites (Δ13Chumans-animals) suggest that millet was predominantly part of the human diet and not regularly used as animal fodder (Fig. S2A, B). The variability in animal isotopic values observed across individual sites suggests differences in livestock foddering practices, such as the proportion of forest versus open pastures or varied winter fodder types, reflecting considerable site-to-site variation. No clear geographical trend emerges at the regional scale, though the available data for Moravia remain limited.

In summary, carbon isotopic values from Iron Age cemeteries demonstrate considerable inter-site variability. Although there is an overall trend towards higher δ13C values in the later phases (LT B2b—LT C1, before mid-third century BCE), especially in the territory of Moravia, in terms of individual sites, a significant shift occurs only at part of the largest cemeteries. The most pronounced positive δ13C shift occurs predominantly among the common group (Group 2), reflecting higher dietary variability within this social cohort. For both rich grave goods groups, δ13C values predominantly remain below the –18‰ (VPDB) threshold, along with still a substantial portion of Group 2, indicating that, despite the visibly increased role of millet, C₃ plants continued to form the dietary staple for the studied populations.

Discussion

Socio-economic and cultural implications of shifting millet reliance

In this study, the positive carbon isotopic shift, observed for the third century BCE, not accompanied by evidence of statistically significant supra-regional increase in millet cultivation, suggests intensified millet consumption without a corresponding systematic agricultural investment in the crop. This scenario is plausible under several conditions discussed below, highlighting the necessity for critical analysis when integrating multi-proxy data. The irregularity of the isotopic dietary shift observed among the studied sites also necessitates a more nuanced interpretative approach. Recent studies have often linked positive shifts in δ13C values to climatic fluctuations, particularly periods of drought37,38,57,67,69. While several studies support this interpretation38,57, climate models for the late Iron Age period instead indicate cooling conditions and increased precipitation70. Furthermore, most Iron Age settlements were established on humus-rich chernozem soils within the traditional agricultural zones of Central Europe71, which are more challenging to cultivate using traditional tillage methods during drought conditions. This evidence, considered alongside the broader societal transformations of the third century BCE, suggests that rather than coping with drought, Iron Age communities adopted multiple agricultural strategies aimed at enhancing overall productivity, diversity, and resilience. More importantly, the results of the archaeobotanical analysis show that this fluctuation occurred without millet ever overtaking the other staple crops in cultivation, suggesting that it was a stable part of the staple crops spectrum that only provided resilience if needed. Therefore, even if millet functioned exclusively as a buffer crop during periods of unfavourable climatic conditions, its use in the Iron Age does not appear to have been a uniform or widespread adaptive strategy.

An important factor for the interpretation of the temporary surge of millet consumption is therefore its selective and variable character. It underscores the capacity of Iron Age communities to flexibly adjust agricultural strategies according to emerging socio-economic needs on a more complex level; a dynamic relevant for understanding adaptive governance in contemporary production systems72. Where the positive isotopic shift was recorded within the analysed area, it was typically associated with larger communities (Fig. 5); only in Moravia is the shift detectable more broadly across the region. The larger cemeteries are predominantly situated within the most agriculturally productive zones, where the industrial agglomerations were established during the course of the third century BCE. The concentration of population within these settlements, coupled with growing economic specialisation, and proximity to long-distance communication networks, required corresponding developments in agriculture aimed at increasing yields and ensuring a consistent surplus. Evidence for intensification in the animal husbandry with focus on working animals in the case of agglomerations in the Middle Danube region12, implies an expansion of cultivation areas within their hinterlands to enhance crop production. Under these conditions, millet would have been particularly attractive due to its agronomic characteristics, including flexibility, short growing season, high yield potential, and suitability for long-term storage17,49. Its fast maturation and summer planting cycle meant it could be slotted into agricultural schedules with relative ease and could serve as a catch-crop or a second harvest or be cultivated on soils and in microclimates less suitable for barley and wheat. These attributes offered farming communities considerable flexibility, enabling them to adapt effectively to fluctuating agricultural needs. Furthermore, millet cultivation may represent a deliberate strategy of agricultural intensification—broadening crop diversity and reinforcing the role of spring-sown cereals alongside traditional staples—to support growing populations in nucleated settlements without requiring a complete shift in dietary practices. This interpretation is reinforced by archaeobotanical evidence from several middle period agglomerations in Western-Central Europe, where millet ranks second in importance to barley10,13,73, a position also similar to ubiquity ratios, MNI, and Representativeness index of millet from contemporaneous Moravian sites (Figs. S7, S11). Conversely, settlements outside these zones may have continued cultivating a staple crop spectrum adapted to local conditions, with variable emphasis on individual crops and local livestock farming. These data collectively point towards substantial flexibility and autonomy among farming communities, each employing subsistence strategies shaped by specific local contexts and preferences. Only during the third century BCE, within emerging agglomerations where economic practices likely became more structured, does the positive δ13C shift appear more systematic—particularly in Moravia, which by this time had become integrated into the wider network of economically progressive Middle Danube zone. In contrast, the adoption of millet-enhanced dietary patterns appears generally more irregular in Bohemia, as visible also in the millet ubiquity ratios and MNI in the archaeobotanical data (Fig. S7, S11). This variability may reflect the peripheral position of Bohemian communities relative to agricultural innovations emerging in the Middle Danube region, with only a few larger sites being integrated into this emerging network.

Another factor shaping variability in dietary patterns is the role of social structures, cultural preferences, and possibly mobility from regions where millet consumption was more common. Previous studies document intensive interaction between the study area and southeastern regions during the third century BCE, involving shared trade networks, technological exchange, and cultural influences5. Demonstrating the presence of direct migrants from these areas, however, requires additional evidence. Strontium isotope data from the Czech Republic, where available alongside dietary isotopes68,74,75, show no consistent link between higher δ13C values with 87Sr/86Sr ratios typical of the Carpathian Basin76. The overall similarity of strontium signatures across loess-dominated lowland zones in Central Europe77 complicates precise identification of migrants and limits the ability to directly attribute dietary shifts to mobility. This is supported by the absence of consistent correlations between δ13C outliers and strontium isotope values. Instead, the variability appears to reflect local dynamics. Most δ13C outliers belong to biological females or individuals from the “common population” (Group 2), whereas groups 1A and 1B are rarely represented (Fig. S21, S22). This suggests that dietary change was primarily an endogenous process affecting the broader population. The two other groups, while following the general trend, retained more conservative and homogeneous dietary profiles, showing only modest increases in δ13C values and continued preference for C₃ plants (Fig. S23). This may reflect socio-cultural preferences among some, perhaps better-provisioned, households, which allowed them to sustain distinct dietary standards regardless of broader shifts in the available food spectrum.

It is noteworthy that during the late La Tène phase (180 BCE – 0 BCE/CE), millet’s role within the staple crop spectrum markedly declined (Fig. S8, S12). This suggests that millet cultivation was no longer a key component of the systematic agricultural strategies adopted by the emerging socio-economic centres in Central Europe from the second century BCE onward. One possible explanation is that oppida—often established at higher elevations—required different cultivation strategies adapted to local environmental conditions. However, archaeobotanical evidence also shows a decline in millet at lowland rural settlements, indicating that its earlier prominence may have been specific to lowland agglomerations. Following the abandonment of these sites, rural communities appear to have reverted to locally adapted agricultural practices, while the oppida developed distinct cultivation regimes, often placing greater emphasis on livestock-based economies and more resilient crop types78,79,80,81,82,83.

All these findings was not a universally adopted agricultural innovation, but rather a preferential dietary choice shaped by geography, economic strategy, cultural tradition, and social organisation. This interpretation highlights how prehistoric communities either adapted or conservatively maintained their foodways in response to environmental and socio-cultural dynamics.

Integrating archaeobotanical and isotopic evidence

Combining archaeobotanical data with stable isotope analysis enables a more comprehensive reconstruction of ancient dietary practices than either approach can achieve alone. Isotopic analysis can detect broad dietary trends (such as the relative importance of C₃ versus C₄ plants or trophic levels) while archaeobotanical evidence provides direct, physical records of specific cultivated plants and their local management. Nevertheless, integrating these two approaches is not always straightforward. Macro-remains typically reflect short-term or site-specific food availability due to their context-dependent preservation. In contrast, stable isotope data from bone collagen represent dietary intake averaged over extended periods of adolescence due to the variable turnover rate of collagen especially in various types of bones84,85,86. Aligning these datasets chronologically and contextually thus presents challenges, particularly when isotopic and archaeobotanical data originate from different sites. Where direct comparisons are not feasible, interpretations should minimise potential errors by carefully selecting assemblages from burial and settlement sites that are as closely matched in time and space as possible. Additionally, developing a detailed understanding of regional archaeological and archaeobotanical patterns is crucial. Therefore, significant attention must be paid to precise dating, classification and contextualisation of the data, enabling meaningful comparisons between individual sites at both local and regional scales. As highlighted in the results, dietary patterns identified from isotopic evidence varied considerably among individual sites; thus, comparisons with archaeobotanical data at a local scale were essential to determine whether millet proportions within local crop spectra diverged from broader, supra-regional trends. However, even when working with datasets compatible chronologically and spatially, the presence of particular plant species within a crop assemblage does not have to necessarily reflect their dietary importance. Beyond simple quantification of macro remains, it is therefore essential to incorporate additional measures such as ‘ubiquity ‘ and the ‘Representativeness index ‘. Only after establishing the ‘economic importance’ of a plant taxon through rigorous archaeobotanical methods, the potential discrepancies between archaeobotanical and isotopic evidence can be approached from the perspective of the true dietary significance of that plant within the studied human populations.

Conclusion

This study demonstrates that the increased reliance on millet during the third century BCE in La Tène Central Europe was a context-dependent, socially stratified, and regionally variable process. Isotopic and archaeobotanical evidence reveal that this dietary shift was not uniform across time or space: the most significant increase occurred in lowland agglomerations during the mid-La Tène period, while millet remained marginal in upland oppida and declined again after 180 BCE. It was also revealed that while millet consumption intensified among certain segments of the population, particularly individuals with simpler grave goods, those buried with warrior or rich female attire maintained more conservative diet with limited millet intake. These developments likely reflect the emergence of nucleated, industrial-style agglomerations, accompanied by population growth and a resulting need to adapt agricultural production towards more efficient and sustainable practices. The observed patterns, both systemic and locally variable, highlight the value of multi-proxy approaches for understanding human adaptability. Only by informed integrating archaeobotanical and isotopic evidence can the nuanced role of millet in Iron Age diets be fully understood.

More broadly, the findings suggest that past dietary practices were shaped not only by environmental conditions, but also by cultural preferences and technological change, resulting in locally adapted strategies. This interplay between resilience, adaptability, and social identity offers meaningful comparative insights for modern societies striving toward more flexible and sustainable food systems.

Methods

Both the isotopic and archaeobotanical analyses followed a chronological and geographical structure, dividing the study area into three main regions (Bohemia, Moravia, Slovakia) and three chronological periods. To mitigate potential biases in regional relative dating, used by archaeologists, when defining phases, absolute date ranges were used for the chronology. This approach allowed for a more accurate consideration of the site’s beginning and end, particularly for settlements where the precise duration of occupation is often difficult to determine. Isotopic data are available only for certain parts of the early and middle phases (400–180 BCE) and are accordingly grouped into two phases corresponding to the fourth century and the third century BCE, respectively. In addition to the division into the main chronological periods, the datasets and statistical outputs also incorporate relative chronological phases as defined in the Central European Iron Age chronology allowing for more detailed chronological overview87. The complete chronological framework used for this study is as follows:

  • Botanical data:

  • Early phase (500–330 BCE)

  • Middle phase (330–180 BCE)

  • Late phase (180 BCE–0 BCE/CE)

  • Isotopic data:

  • Early phase (400–300/290): LT B1, LT B2a

  • Middle phase (300/290–200/180 BCE): LT B2b, LTB2b-C1, LT C1

The phases in the archaeobotanical dataset are extended beyond chronological delimitation of periods, because the settlements are dated in longer chronological intervals.

Botanical analysis

The archaeobotanical data were collected from both published and unpublished sources and originated from 99 archaeological sites providing 143 analytical units including sites or site-related chronological-phase assemblages (Table S3). The archaeobotanical dataset comprises all available data on cultivated and wild plant seeds accessible to the authors from Slovakia (Table S4). For Moravia and Bohemia, both published and unpublished archaeobotanical data from the Czech Republic, compiled in the Czech Archaeobotanical Database, were collected. Plant macro-remain data were assembled for all cultivated plants (cereals, pulses, and oil/fibre crops), potentially edible wild fruits and nuts, as well as the total number of weed seeds. For each taxon, the total number of identified specimens (NISP) of grains, seeds, and chaff, along with the number of samples in which the taxon was recorded, were documented for each site and chronological phase (Tables S5, S6). Grain and chaff data were combined to calculate the MNI (minimum number of individuals) of potential grains per site. Ratios of grain to chaff components were applied as follows: einkorn (1:2), emmer, spelt, and timopheevii (2:2), naked wheat and barley (3:1), rye (2:1), and oats (1:1) (Table S7). Additionally, presence/absence and ´ubiquity´ ratios (the number of samples in which a taxon occurred divided by the total number of samples from a site) were used to mitigate biases from differing methodologies, researchers, or geographical contexts. To address data variability, the data were standardised by calculating taxon percentages relative to the total (MNI values) of identified cereal finds, excluding indeterminate grain fragments. For sites with fewer than 5 samples, modelled ubiquity data (calculated in two ANOVA runs) were employed to address small sample size constraints. For millet, values ranging from zero to the phase average were replaced with the lower quartile, while values above the average were replaced with the upper quartile of the phase value (Table S8). Sites with fewer than 50/30 finds (depending on percentage calculations) or those with broad chronological ranges were excluded when evaluating ubiquity ratios.

The statistical evaluation aimed to identify chronological and spatial patterns of millet in the study area. Archaeobotanical data were analysed at the site or site´s chronological phase level. A Detrended Correspondence Analysis (DCA) was conducted to investigate variations in cereal crop composition across sites from different chronological periods or regions. This analysis utilised two standardised matrices of cereal crops: (a) percentages and (b) ubiquities (ratios) per site. DCA was performed using Canoco for Windows88. Differences in central values and distributions among groups were tested using the non-parametric Kruskal–Wallis test and the parametric one-way ANOVA. Additionally, the Representativeness index (RI) was calculated, which assesses both the abundance and dominance of individual taxa in relation to the total macrofossil count and the number of analysed samples from each site. The RI was calculated using the approach of Stika and Heiss89. By combining frequency, ubiquity, and RI measures, this approach provides the most reliable approximation of the economic importance of millet within the given time period and study area.

The cartographical distribution of the data (ArcGIS Pro, ESRI Inc.) includes information on both the percentage and ubiquity of millet at the analysed sites (Fig. S5S12). The graphic outputs utilise pie charts to represent the proportion of millet in crop spectrum per site. The size of the pie charts is normalised according to the total cereal counts at each site for percentage charts and number of samples for ubiquity charts.

Isotope analysis

Data on the isotopic composition of collagen carbon and nitrogen from La Tène cemeteries in the region, along with archaeological contextualisation regarding dating and grave goods, were compiled into a single dataset containing 429 samples from 16 burial sites, with available δ13C and δ15N collagen values and archaeological contextualisation. The data were compiled from the available publications and also include 79 new isotopic data from the Moravian sites of Blučina, Nechvalín, Lovčičky, and Brno Maloměřice that were recently analysed (Table S9, S12, SI_protocol). Where available, the isotopic data from animals were included in the dataset; this information was available for 9 sites and included 87 samples (Table S10). Chronological variations were examined between the fourth and third centuries BCE (the latter period being of particular interest for identifying socio-economic transformations) as well as across relative chronological phases.

The statistical analysis of the isotopic dataset (SI_analysis) focused on identifying chronological, social, and geographical patterns in the study region. Individual sites were analysed separately, examining carbon and nitrogen isotopic compositions across chronological phases and associating them with defined main social categories, grouped as “warrior equipment” (group 1A), “rich female attires” (group 1B), and “general population” (group 2). To account for environmental and dietary variations among sites and regions, isotopic data from animals were also collected where possible and compared with human data.

Statistical differences among group central values and distributions were tested using non-parametric tests (Mann–Whitney U, Kruskal–Wallis) as they are more efficient for analysing subsets of data and non-normal distributions. The resulting p-value was subsequently assessed in conjunction with data visualisation. Individual groups were also compared using kernel density estimation (KDE). KDE is a non-parametric method for estimating the probability density function of continuous variables, providing a smooth, distribution-independent representation of the data. Studies have demonstrated that KDE provides a more precise and flexible method for analysing isotope data90. This approach allows for capturing the potential trends in millet consumption. The cartographical distribution of the data (ArcGIS Pro, ESRI Inc., https://www.arcgis.com), used both in the manuscript and supplementary files (Fig. S3S12) includes spatial information on recorded isotopic shifts. If not stated explicitly, for all the background maps the World Topographic Map by © Esri and NASA, NGA, USGS, GUGiK, ŠOP SR, Esri TomTom, Garmin, FAO, METI/NASA, was used.