Abstract

The ability to detect and control fuel isotopic content down to a 1% concentration level is greatly important for the upcoming JET DTE2 campaign, as well as its associated TT and DD phases. A reduction of H minority concentration even from 2% down to 1% is shown here to have significant impact on the effectiveness of ion cyclotron range of frequencies core heating, while the ability to maintain T or D concentration at or below 1% is critical to limiting fusion neutron generation in the DD and TT phases, correspondingly. The sub-divertor measurement of (global) isotopic concentration, based on Penning-activated optical spectroscopy, can deliver minimally this 1% detection for DTE2 as long as light collection from the Penning emission can be optimized and gradual window transmission deterioration can be minimized. This is simulated with a statistical analysis developed to understand the uncertainty sources in the JET DTE1 data, as well as to guide the optimization of an upgraded, fuel-isotopic content (and helium-ash concentration) gas analysis system for the JET divertor in preparation for DTE2. While this random error can be reduced to allow measurement substantially below 1% concentration, analysis also shows a systematic error of up to 1% understood to be due to plasma–surface interactions in the Penning excitation, suggesting that 1% may still be the low-end limit for the sub-divertor measurement, unless a Penning-source conditioning approach is also developed.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call