Five experienced observers assessed 100 antepartum cardiotocograms from 100 pregnant women by means of three different systems: two commonly used scoring systems and a close look at a 1-minute recording period (window). Variation between observers (interobserver) and within one observer (intraobserver) was determined by calculating the weighted kappa coefficients. Weighted kappa coefficients for interobserver variation in the Visser/Huisjes score and the Fischer score were 0.41 and 0.37, respectively. Weighted kappa coefficients for interobserver variation in a close look at a 1-minute window ranged from 0.09 to 0.69. Thus, a low level of agreement between observers was shown to exist for all three systems tested. The level of intraobserver agreement was shown to be much higher for virtually all three systems. Since the level of interobserver agreement is low, the results of different studies that describe visually assessed antepartum cardiotocograms by means of these scoring systems or 1-minute windows are not comparable, and reproducibility will be low.