The interaction between turbulence in a minimal supersonic channel and radiative heat transfer is studied using large-eddy simulation. The working fluid is pure water vapour with temperature-dependent specific heats and molecular transport coefficients. Its line spectra properties are represented with a statistical narrow-band correlated-k model. A grey gas model is also tested. The parallel no-slip channel walls are treated as black surfaces concerning thermal radiation and are kept at a constant temperature of 1000 K. Simulations have been performed for different optical thicknesses (based on the Planck mean absorption coefficient) and different Mach numbers. Results for the mean flow variables, Reynolds stresses and certain terms of their transport equations indicate that thermal radiation effects counteract compressibility (Mach number) effects. An analysis of the total energy balance reveals the importance of radiative heat transfer, compared to the turbulent and mean molecular heat transport.