new

Get trending papers in your email inbox!

Subscribe

Daily Papers

byAK and the research community

Mar 13

Investigation of W-SiC compositionally graded films as a divertor material

W-SiC composite material is a promising plasma-facing material candidate alternative to pure W due to the low neutron activation, low impurity radiation, and low tritium diffusivity of SiC while leveraging the high erosion resistance of the W armor. Additionally, W and SiC have high thermomechanical compatibility given their similar thermal expansion rates. The present study addresses the synthesis and performance of compositionally graded W-SiC films fabricated by pulsed-DC magnetron sputtering. Compositional gradients were characterized using transmission electron microscopy (TEM) and energy-dispersive X-ray spectroscopy (EDS), and crystallographic information was obtained using electron diffraction and X-ray diffraction (XRD). Samples were exposed to L-mode deuterium plasma discharges in the DIII-D tokamak using the Divertor Material Evaluation System (DiMES). Post-mortem characterizations were performed using scanning electron microscopy (SEM) and XRD. Electron diffraction and XRD showed that the compositionally graded W-SiC films were composed of polycrystalline W and amorphous SiC with amorphous W+SiC interlayers. No macroscopic delamination or microstructural changes were observed under mild exposure conditions. This study serves as a preliminary examination of W-SiC compositionally graded composites as a potential candidate divertor material in future tokamak devices.

  • 16 authors
·
Aug 30, 2023

DeepDistill: Enhancing LLM Reasoning Capabilities via Large-Scale Difficulty-Graded Data Training

Although large language models (LLMs) have recently achieved remarkable performance on various complex reasoning benchmarks, the academic community still lacks an in-depth understanding of base model training processes and data quality. To address this, we construct a large-scale, difficulty-graded reasoning dataset containing approximately 3.34 million unique queries of varying difficulty levels and about 40 million distilled responses generated by multiple models over several passes. Leveraging pass rate and Coefficient of Variation (CV), we precisely select the most valuable training data to enhance reasoning capability. Notably, we observe a training pattern shift, indicating that reasoning-focused training based on base models requires higher learning rates for effective training. Using this carefully selected data, we significantly improve the reasoning capabilities of the base model, achieving a pass rate of 79.2\% on the AIME2024 mathematical reasoning benchmark. This result surpasses most current distilled models and closely approaches state-of-the-art performance. We provide detailed descriptions of our data processing, difficulty assessment, and training methodology, and have publicly released all datasets and methods to promote rapid progress in open-source long-reasoning LLMs. The dataset is available at: https://huggingface.co/datasets/a-m-team/AM-DeepSeek-Distilled-40M

  • 8 authors
·
Apr 24, 2025