Platform Tour
The IOMETE Platform interface is composed of the following sections:
-
User Menu: - Use this menu to switch themes, access the settings menu, and log out.
-
Navigation Menu: - Navigate through different Platform areas including Lakehouses, SQL Editor, Spark jobs, Data catalog and more.
-
Help Menu: - Access this menu to redirect to Community & Support and Documentation for assistance.
-
Content: - When a Platform menu item is selected, the associated content is displayed in the content pane.
Engine
Lakehouses page
A virtual lakehouse is a cluster of compute resources that provide the required resources, such as CPU, memory to perform the querying processing. Before using the SQL Editor, you should create a lakehouse.
For more details, see Virtual Lakehouses
Spark Connect page
Spark Connect, introduced in Apache Spark > 3.4, provides a decoupled client-server architecture for remote connectivity to Spark clusters using the DataFrame API. This separation allows Spark and its open ecosystem to be leveraged from various environments, including modern data applications, IDEs, Notebooks, and programming languages.
For more details, see Spark Connect
Jupyter Kernels page
Jupyter Notebook is a powerful tool for interactive data exploration and prototyping. By connecting to IOMETE's Jupyter Gateway you will be able to explore and analyze data stored in IOMETE's data lake directly from your local environment.
For more details, see Starting with Notebook
SQL
SQL Editor page
The main components of the SQL Editor are highlighted below:
- Worksheets offer a simple method for you to write SQL queries, execute the queries, and see the results.
- Database explorer is used to explore your database objects, which includes namespaces, tables, views, and their columns (even complex columns), and partitions.
- Lakehouse and Namespace selection are required to execute SQL queries.
- SQL Editor is used to write SQL queries using autocomplete feature. IOMETE tracks database objects and suggests them as autocomplete options
- Query result view provide executed queries results. You can filter each column in the results and toggle on Charts to visualize the results as graphs.
For more details, see SQL Editor
Query History page
Query Histories track each query you run, adding it as a new item in the SQL history. You can revisit previous queries to view their results and reopen them in the active worksheet.
For more details, see Query History
Applications
Spark Applications page
The Spark Applications page provides a centralized view of all Spark applications across various jobs, visualized with interactive charts. Users can filter applications by time range, user, status, or tags to efficiently monitor and analyze performance. This comprehensive view aids in tracking application progress, identifying issues, and optimizing resource usage.
Job Templates page
IOMETE offers PySpark quickstart template for AWS or PySpark quickstart template for GCP helping you kickstart your first Spark Job. Follow the README instructions to use the template, which includes a sample job reading a CSV file from an S3 bucket, applying transformations, and writing the output to an Iceberg Table. Use it as a starting point for your custom jobs.
For more details, see Spark job
Governance
Search
The data catalog interface offers metadata for all tables. You can easily filter by Schemas, Table type, Provider and Tags, and use the search function to find what you need. Additionally, you have the option to bookmark tables for quick access.
Data Explorer page
The Catalog Explorer page allows you to view and manage all catalogs, namespaces, and tables with ease. You can register or snapshot tables and access detailed information, including table details, partition information, column data, and snapshots.
Data security page
Data security at IOMETE is maintained through access control using a user interface for consistent administration. Security admins set policies for databases, tables, and columns, managing permissions for groups or users. IOMETE offers the following types of data security.
- Access Policy policy comprises rules and permissions that specify who can access specific resources and the actions they can take after gaining access.
- Data masking is a method to hide sensitive information by changing it into fake data while keeping its appearance intact. It safeguards private details during testing or sharing.
- Row-level filtering limits who can see certain rows in a database.
For more details, see Data security overview