Inference4j 0.9.0 Released: Summarization, Translation, Text2SQL Support

Inference4j 0.9.0 released, Summarization, Translation, Text2SQL support Summarization, translation, grammar correction, text-to-SQL — all running locally on the JVM with no Python, no microservices, no GPU required.   try (var summarizer = BartSummarizer.distilBartCnn().build()) {       String summary = summarizer.summarize(longArticle);       System.out.println(summary);   }   What's new in 0.9.0:   - Seq2seq inference engine with KV cache — the foundation for all encoder-decoder models   - 5 new task wrappers: FlanT5, BART summarizer, MarianMT translator, CoEdit grammar corrector, T5 text-to-SQL   - Unigram (SentencePiece) tokenizer with Viterbi decoding   - ByteBuffer pooling for lower GC pressure   - Download progress tracking Text2SQL models allows developers to use natural language and translate into real SQL queries to the SQL DB. This one I think may be a popular one for java devs.   That brings inference4j to 19 task wrappers supporting 30+ models across vision, audio, NLP, and multimodal — all through type-safe, builder-pattern APIs that feel like writing normal Java.   The goal hasn't changed: make on-device AI inference a first-class citizen in the Java ecosystem. No ONNX tensors, no JNI juggling — just pick a model, call a domain method.   What's coming next:   - 0.10.0 — Named Entity Recognition + more embedding models   - 0.11.0 — TikToken tokenizer, unlocking Llama 3.2 and other modern LLMs   - 0.12.0 — Text-to-speech pipeline via Piper — local voice synthesis on the JVM Docs: https://lnkd.in/e63MXwQN   GitHub: https://lnkd.in/eEUESRkq   #Java #AI #MachineLearning #ONNX #OpenSource

To view or add a comment, sign in

Explore content categories