Skip to main content

Marketplace Jobs

Latest Versions

NameImageVersionLink
Data Compactioniomete_data_compaction1.2.12Open ↗
File Streamingiomete_file_streaming0.3.0Open ↗
Catalog Synciom-catalog-sync4.3.5Open ↗
MySQL Synciomete_mysql_sync3.0.0Open ↗
Kafka Iceberg Streamkafka-iceberg-stream1.2.0Open ↗

Recent Releases

January 7, 2026

Catalog Sync Job - v4.3.5

⚡ Improvements

Added new metrics to track Iceberg table and database sizes including all snapshots:

  • Number of files for an Iceberg table including all snapshots (so showing true number of files for a table if they look into their storage)

  • Total size of an Iceberg table including all snapshots

  • Total DB/schema size including all snapshots

    New Metrics | IOMETENew Metrics | IOMETE

These metrics are new columns in the existing tables in iomete_catalog_db:

  • table_metadata: total_table_nums_files, total_table_size_in_bytes
  • schema_metadata: total_db_size_in_bytes
    Table Metadata | IOMETETable Metadata | IOMETE
    Schema Metadata | IOMETESchema Metadata | IOMETE
January 7, 2026

Useful SQL tools - v1.0.0

🚀 New Features

This is not a Docker image release, but rather a release of a set of useful SQL tools for managing and querying Iceberg tables in IOMETE. See sql scripts in the GitHub repository: iomete/iomete-marketplace-jobs/sql-tools

Currently, these tools include:

  • Creation of iomete_spark_audit_external_table to audit Spark jobs. From this table, one can query

    • run time (using eventTime column)
    • run date (using day column)
    • user (using user column)
    • job/query id (using eventId column)
  • Creation of view for Daily jobs counts for a particular day of interest (day has to be hardcoded). This will create table stats for that day like below:

    Schema Metadata | IOMETESchema Metadata | IOMETE
  • Query that filters Spark jobs for multiple days. This query can last for some time depending on how much data accumulated on each day, and how many days is being queried:

    Schema Metadata | IOMETESchema Metadata | IOMETE
  • Query that helps track IOMete Database Objects Definitions (DDL) Change Tracking and Audit Capability

November 3, 2025

Data Compaction Job - v1.2.12

🚀 New Features

  • Implemented table-level locking system to prevent concurrent compaction operations on the same table
  • Config examples:
        lock {
    enabled = true
    ttl_seconds = 172800 # 48 hours (covers 1-day worst-case + buffer)
    }
October 21, 2025

Data Compaction Job - v1.2.11

🚀 New Features

  • Time-based Snapshot Expiration:
    • Added support to remove Iceberg snapshots older than a set number of days using expire_snapshot.older_than_days.

    • Works together with retain_last. If both are set, we keep snapshots that match either rule. We always keep at least 1 snapshot.

    • Config examples:

      // Remove snapshots older than 7 days (keep at least 1)
      expire_snapshot: { older_than_days: 7 }

      // Keep last 3 OR anything newer than 7 days (whichever is more)
      expire_snapshot: { retain_last: 3, older_than_days: 7 }

      // Table-specific override
      table_overrides: {
      production.critical_table: {
      expire_snapshot: { retain_last: 10, older_than_days: 30 }
      }
      }
    • Retention rules:

      ConfigurationBehavior
      None specifiedKeeps 1 snapshot (default)
      Only retain_lastKeeps the last N snapshots
      Only older_than_daysRemoves snapshots older than N days (minimum 1 snapshot always kept)
      Both specifiedKeeps snapshots matching EITHER condition (maximum retention)

⚡ Improvements

  • Selective Operation Execution: Added support for fine-grained control over compaction operations through an enabled flag at both global and table-specific levels. You can now selectively enable or disable specific operations.

🐛 Bug Fixes

  • Table Name Resolution:
    • Fixed an issue where table names provided without a database prefix (e.g., my_table instead of db.my_table) in table_include, table_exclude, or table_overrides would incorrectly run on the entire database.
    • The job now correctly resolves such tables using the databases parameter, ensuring targeted execution.
  • Rewrite Data Files with WHERE Filter:
    • Fixed an issue where the where filter failed due to incorrect string handling.
    • The job now correctly wraps SQL expressions, allowing compaction with both static and dynamic date filters.
    • Config examples:
        rewrite_data_files: {
      // Static date filter
      where: "date >= '2025-01-01'"

      // Dynamic filters (recommended - no manual date updates needed)
      // where: "date <= CURRENT_DATE - 30" // Data older than 30 days
      // where: "event_time <= CURRENT_TIMESTAMP - INTERVAL 1 DAY" // Data older than 1 day
      }