Skip to content

Releases: keras-team/keras-hub

v0.26.0

06 Feb 06:55
12bf85a

Choose a tag to compare

New Models

  • Translate Gemma: A multimodal variant of the Gemma 3 model fine-tuned for high-quality machine translation across 55 languages and capable to translate many more languages both directions, supporting both text and image inputs.
  • SAM3 (Segment Anything Model 3): A next-generation computer vision model that introduces Promptable Concept Segmentation (PCS), allowing for precise object and concept segmentation through text or visual prompts.
  • Qwen 2.5 Coder: A code-specialized version of the Qwen 2.5 series, optimized for programming tasks, debugging, and code generation across a wide variety of programming languages.
  • Qwen 2.5 Math: A specialized variant of the Qwen 2.5 family designed for advanced mathematical reasoning, capable of solving complex problems with high precision.
  • Qwen 3 Coder: An advanced coding Mixture-of-Experts model built on the Qwen 3 MoE architecture, delivering exceptional performance across both programming benchmarks and agentic tasks.
  • RWKV7: A high-performance, fully recurrent (100% RNN) architecture featuring linear-time complexity and constant-space inference. By eliminating the need for a KV-cache and standard attention mechanisms.

Export to Safetensors

  • Added Gemma3 Text models support for Safetensor export.
  • Added Qwen Text models support for Safetensor export.

New Features

  • Hugging Face Porting Script: Added an automated script to port any text-only decoder LLM from Hugging Face to the Keras Hub repository.
  • AWQ Support: Added support for Activation-aware Weight Quantization (AWQ).

Bug Fixes and Improvements

  • Python 3.13 Compatibility: Made tensorflow-text an optional dependency to ensure compatibility with Python 3.13.
  • Masking: Fixed masking issues in TokenAndPositionEmbedding and improved compatibility with JAX.
  • Security: Fixed a safe mode bypass vulnerability in tokenizers.
  • Numerical Stability: Fixed a float16 overflow issue in Gemma 3.

Contributors

We would like to thank our contributors for this release:
@Amitavoo, @amitsrivastava78, @divyashreepathihalli, @gaurides, @hertschuh, @james77777778, @jaytiwarihub, @JyotinderSingh, @kharshith-k, @LakshmiKalaKadali, @laxmareddyp, @mattdangerw, @nikolasavic3, @pass-lin, @pctablet505, @sachinprasadhs, @shashaka.

Full Changelog: v0.25.1...v0.26.0

v0.26.0.dev0

06 Feb 00:54
6743287

Choose a tag to compare

v0.26.0.dev0 Pre-release
Pre-release

What's Changed

New Contributors

Full Changelog: v0.25.1...v0.26.0.dev0

v0.25.1

13 Jan 21:18
4a41006

Choose a tag to compare

What's Changed

  • Fix float16 overflow in Gemma3 by addressing precision-related instabilities.

Full Changelog: v0.25.0...v0.25.1

v0.25.0

18 Dec 16:33
f1da303

Choose a tag to compare

Summary:

New Models:

We've integrated new open-weight models to expand the capabilities of KerasHub, featuring specialized tools for function calling and safety, as well as high-performance open-source reasoning models:

  • FunctionGemma: We have added support for FunctionGemma, a lightweight model from Google built on the Gemma 3 270M architecture. Designed specifically for text-only function calling, this model is optimized for single-turn scenarios and deployment in resource-constrained environments.
  • GPT OSS: We have integrated OpenAI’s gpt-oss family, including the 20B and 120B parameter variants. These models utilize a Mixture-of-Experts (MoE) architecture with a 128k token context window, optimized for STEM, coding, and general reasoning tasks.
  • GPT OSS Safeguard: A new open-weight safety reasoning model from OpenAI. Built upon the GPT OSS architecture, it enables adaptable content classification and input-output filtering based on custom safety policies.

What's Changed

Full Changelog: v0.24.0...v0.25.0

v0.25.0.dev0

18 Dec 08:42
63a87ed

Choose a tag to compare

v0.25.0.dev0 Pre-release
Pre-release

What's Changed

Full Changelog: v0.24.0...v0.25.0.dev0

v0.24.0

05 Dec 05:31
6a32eea

Choose a tag to compare

Summary:

New Models:

We've integrated new models and presets to expand the capabilities of KerasHub:

  • DINOv3: We have added the DINOv3 model architecture and registered its corresponding presets.
  • MedGemma & MedSigLIP: New presets have been registered for MedGemma and MedSigLIP, bringing specialized capabilities for medical domain tasks.
  • Qwen3 Embeddings: We have registered embedding presets for the Qwen3 model family.

Improvements & Enhancements

This update includes infrastructure improvements and fixes:

  • GPTQ Quantization Hooks: Added get_quantization_layer_structure hooks to facilitate GPTQ quantization support.
  • TensorFlow Compatibility: Fixed tensorflow-text imports to ensure they do not break core TensorFlow functionality.
  • Gemini CLI Workflow: Introduced a new workflow to support co-working with the Gemini CLI.

What's Changed

New Contributors

Full Changelog: v0.23.0...v0.24.0

v0.24.0.dev0

04 Dec 21:42
466acdc

Choose a tag to compare

v0.24.0.dev0 Pre-release
Pre-release

What's Changed

New Contributors

Full Changelog: v0.23.0...v0.24.0.dev0

v0.23.0

21 Oct 17:34
2dc4d39

Choose a tag to compare

Summary:

New Models:

We've integrated a range of cutting-edge models, each designed to tackle specific challenges in their respective domains:

  • Cell2Sentence: A single-cell, biology-aware model built on the Gemma-2 architecture, designed to interpret complex biological data.

  • T5Gemma: A new encoder-decoder model, ideal for sequence-to-sequence tasks like translation and summarization.

  • PARSeq: An end-to-end, ViT-based model for scene text recognition (STR), excelling at reading text in natural images.

  • D-FINE: A high-performance, real-time object detection model.

  • DepthAnythingV2: A monocular depth estimation (MDE) model trained on a combination of synthetic labeled data and real-world unlabeled images.

  • Qwen3 Moe: The largest language model in the Qwen series, utilizing a Mixture-of-Experts (MoE) architecture for enhanced performance and efficiency.

  • MobileNetV5: A state-of-the-art vision encoder specifically designed for high-efficiency AI on edge devices.

  • SmolLM3: A compact yet powerful language model excelling in reasoning, long-context understanding, and multilingual capabilities.

Improvements & Enhancements

This update also includes several key improvements to enhance the platform's stability, compatibility, and flexibility:

  • export_to_transformers: You can now export trainable models, tokenizers, and configurations directly into the Hugging Face Transformers format using export_to_transformers. This feature is currently available for Gemma models, with support for more architectures coming soon.
  • OpenVINO Backend Support: We've integrated OpenVINO inference support, enabling optimized inference for Mistral, Gemma, and GPT-2 models.
  • Bidirectional Attention Mask: Gemma models now support a bidirectional attention mask, enabling more effective fine-tuning on tasks that require understanding the full context of a sequence.
  • CLIP & SD3 Model Refactor: The CLIP and Stable Diffusion 3 models have been refactored to improve numerical stability. Updated checkpoints are now available to ensure seamless and reliable performance.

What's Changed

New Contributors

Full Changelog: v0.22.2...v0.23.0

v0.23.0.dev0

20 Oct 23:35
4de2ff6

Choose a tag to compare

v0.23.0.dev0 Pre-release
Pre-release

What's Changed

New Contributors

Full Changelog: v0.22.2...v0.23.0.dev0

v0.22.2

12 Sep 15:31
f4b648d

Choose a tag to compare

New Model: VaultGemma

VaultGemma is a 1-billion-parameter, 26-layer, text-only decoder model trained with sequence-level differential privacy (DP).
Derived from Gemma 2, its architecture notably drops the norms after the Attention and MLP blocks and uses full attention for all layers, rather than alternating with local sliding attention.
The pretrained model is available with a 1024-token sequence length.

What's Changed

Full Changelog: v0.22.1...v0.22.2