BART Large CNN โ€” ONNX

ONNX export of BART Large CNN (406M parameters) with encoder-decoder architecture and KV cache support.

Fine-tuned for text summarization on the CNN/DailyMail dataset.

Converted for use with inference4j, an inference-only AI library for Java.

Original Source

Usage with inference4j

try (var summarizer = BartSummarizer.bartLargeCnn().build()) {
    System.out.println(summarizer.summarize("Long article text..."));
}

Model Details

Property Value
Architecture BART encoder-decoder (406M parameters, 12 encoder + 12 decoder layers)
Task Text summarization
Training data CNN/DailyMail
Tokenizer BPE (50,265 tokens)
Original framework PyTorch (transformers)
Export method Hugging Face Optimum (encoder-decoder with KV cache)

License

This model is licensed under the Apache License 2.0. Original model by Facebook AI.

Downloads last month
11
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support