Tellius
Tellius 5.5
Tellius 5.5
  • 🚩Getting Started
    • 👋Say Hello to Tellius
      • Glossary
      • Tellius 101
      • Navigating around Tellius
    • ⚡Quick Start Guides
      • Search
      • Vizpads (Explore)
      • Insights (Discover)
    • ✅Best Practices
      • Search
      • Vizpads (Explore)
      • Insights (Discover)
      • Predict
      • Data
    • ⬇️Initial Setup
      • Tellius architecture
      • System requirements
      • Installation steps for Tellius
      • Customizing Tellius
    • Universal Search
    • 🏠Tellius Home Page
    • ❓FAQs
      • Kaiya Conversational AI
      • Data Preparation FAQs
      • Environment FAQs
      • Search FAQs
      • Vizpads FAQs
      • Data Caching FAQs
      • Embedding FAQs
      • Insights FAQs
  • Kaiya
    • ♟️Understanding AI Agents & Agentic Flows
      • Glossary
      • Composer
      • 🗝️Triggering an agentic workflow
      • The art of possible
    • 🤹Kaiya conversational AI
      • Triggering Insights with "Why" questions
      • Mastering Kaiya conversational AI
      • 📒Kaiya Learnings
      • Kaiya Terms of Service
  • 🔍Search
    • 👋Get familiar with our Search interface
    • 🤔Understanding Tellius Search
    • 📍Search Guide
    • 🚀Executing a search query
      • Selecting a Business View
      • Typing a search query
      • Constructing effective search queries
      • Marketshare queries
    • 🔑Analyzing search results
      • Understanding search results
      • Search Inspector
      • Time taken to execute a query
      • Interacting with the resulting chart
    • 📊Know your charts in Tellius
      • Understanding Tellius charts
      • Variations of a chart type
      • Building charts from Configuration pane
      • List of chart-specific fields
      • Adding columns to fields in Configuration pane
      • Absolute and percentage change aggregations
      • Requirements of charts
      • Switching to another chart
      • Formatting charts
      • Advanced Analytics
      • Cumulative line chart
    • 🧑‍🏫Help Tellius learn
    • 🕵️‍♂️Search history
    • 🎙️Voice-driven search
    • 🔴Live Query mode
  • 📈Vizpads (Explore)
    • 🙋Meet Vizpads!
    • 👋Get familiar with our Vizpads
    • #️⃣Measures, dimensions, date columns
    • ✨Creating Vizpads
    • 🌐Applying global filters
      • Filters in multi-BV Vizpads
      • Filters using common columns
    • 📌Applying local filters
    • 📅Date picker in filters
      • Customizing the calendar view
    • ✅Control filters
      • Multi-select list
      • Single-select list
      • Range slider
      • Dropdown list
    • 👁️Actions in View mode
      • Interacting with the charts
      • Exporting tables
    • 📝Actions in Edit mode
      • 🗨️Viz-level actions
      • Copy to Clipboard
    • 🔧Anomaly management for line charts
      • Instance level
      • Vizpad level
      • Chart level
    • ⏳Time taken to load a chart
      • Instance level
      • Vizpad level
      • Chart level
    • ♟️Working with sample datasets
    • 🔁Swapping Business View of charts
      • Swapping only the current Vizpad
      • Swapping multiple objects
      • Configuring the time of swap
    • 🤖Explainable AI charts
  • 💡Insights (Discover)
    • 👋Get familiar with our Insights
    • ❓Understanding the types of Insights
    • 🕵️‍♂️Discovery Insights
      • Impact Calculation for Top Contributors
    • ➕How to create new Insights
      • 🔛Creating Discovery Insight
      • 🔑Creating Key Driver Insights
      • 〰️Creating Trend Insights
      • 👯Creating Comparison Insights
    • 🧮The art of selecting columns for Insights
      • ➡️How to include/exclude columns?
  • 🔢Data
    • 👋Get familiar with our Data module
    • 🥂Connect
    • 🪹Create new datasource
      • Connecting to Oracle database
      • Connecting to MySQL database
      • Connecting to MS SQL database
      • Connecting to Postgres SQL database
      • Connecting to Teradata
      • Connecting to Redshift
        • Access S3 Data with Redshift Spectrum
      • Connecting to Hive
      • Connecting to Azure Blob Storage
      • Connecting to Spark SQL
      • Connecting to generic JDBC
      • Connecting to Salesforce
      • Connecting to Google cloud SQL
        • Connecting to a PostgreSQL cloud SQL instance
        • Connecting to an MSSQL cloud SQL instance
        • Connecting to a MySQL Cloud SQL Instance
      • Connecting to Amazon S3
      • Connecting to Google BigQuery
        • Steps to connect to a Google BigQuery database
      • Connecting to Snowflake
        • OAuth support for Snowflake
        • Integrating Snowflake with Azure AD via OAuth
        • Integrating Snowflake with Okta via OAuth
        • Azure PrivateLink
        • AWS PrivateLink
        • Best practices
      • Connecting to Databricks
      • Connecting to Databricks Delta Lake
      • Connecting to an AlloyDB Cluster
      • Connecting to HDFS
      • Connecting to Looker SQL Interface
      • Loading Excel sheets
      • 🚧Understanding partitioning your data
    • ⏳Time-to-Live (TTL) and Caching
    • 🌷Refreshing a datasource
    • 🪺Managing your datasets
      • Swapping datasources
    • 🐣Preparing your datasets
      • 🤾Actions that can be done on a dataset
      • Data Pipeline
      • SQL code snippets
      • ✍️Writeback window
      • 🧩Editing Prepare → Data
      • Handling null or mismatched values
      • Metadata view
      • List of icons and their actions
        • Functions
        • SQL Transform
        • Python Transform
        • Standard Aggregation
        • Creating Hierarchies
      • Dataset Scripting
      • Fusioning your datasets
      • Scheduling refresh for datasets
    • 🐥Preparing your Business Views
      • 🌟Create a new Business View
      • Creating calculated columns
      • Creating dynamic parameters
      • Scheduling refresh for Business Views
      • Setting up custom calendars
      • Custom Calendars for Live Connections
    • Tellius Engine: Comparison of In-Memory vs. Live Mode
    • User roles and permissions
    • Refresh pipeline
  • Feed
    • 📩What is a Feed in Tellius?
    • ❗Alerts on the detection of anomalies
    • 📥Actions done on a tracking Feed
    • 🖲️Track a new metric
  • Assistant
    • 💁Introducing Tellius Assistant
    • 🎤Voice-based Assistant
    • 💬Interacting with Assistant
    • ↖️Selecting Business View
  • Embedding Tellius
    • What you should know before embedding
    • Embedding URL
      • 📊Embedding Vizpads
        • Apply and delete filters
        • Vizpad-related actionTypes
        • Edit, save, and share a Vizpad
        • Keep, remove, drill sections
        • Adding a Viz to a Vizpad
        • Row-level policy filters
      • 💡Embedding Insights
        • Creating and Viewing Insights
      • 🔎Embedding Search
        • Search query execution
      • Embedding Assistant
      • 🪄Embedding Kaiya
      • Embedding Feed
  • API
    • Insights APIs
    • Search APIs
    • Authentication API (Login API)
  • ✨What's New
    • Release 5.5
    • Release 5.4
      • Patches 5.4.0.1 to 5.4.0.4
      • Patch 5.4.0.5
      • Patch 5.4.1
      • Patches 5.4.1.1 and 5.4.1.2
    • Release 5.3
      • Patch 5.3.1
      • Patch 5.3.2
      • Patch 5.3.3
    • Release 5.2
      • Patch 5.2.1
      • Patch 5.2.2
    • Release 5.1
      • Patch 5.1.1
      • Patch 5.1.2
      • Patch 5.1.3
    • Release 5.0
      • Patch 5.0.1
      • Patch 5.0.2
      • Patch 5.0.3
      • Patch 5.0.4
      • Patch 5.0.5
    • Release 4.3 (Fall 2023)
      • Patch 4.3.1
      • Patch 4.3.2
      • Patch 4.3.3
      • Patch 4.3.4
    • Release 4.2
      • Patch 4.2.1
      • Patch 4.2.2
      • Patch 4.2.3
      • Patch 4.2.4
      • Patch 4.2.5
      • Patch 4.2.6
      • Patch 4.2.7
    • Release 4.1
      • Patch 4.1.1
      • Patch 4.1.2
      • Patch 4.1.3
      • Patch 4.1.4
      • Patch 4.1.5
    • Release 4.0
Powered by GitBook

© 2025 Tellius

On this page
  • Pick Python if:
  • Creating and applying Python code
  • Editing Python code

Was this helpful?

  1. Data
  2. Preparing your datasets
  3. List of icons and their actions

Python Transform

Create, edit, and apply Python code transformations to your dataset

Python (whether PySpark or Pandas) is more flexible for applying complex business rules, iterative or row-level manipulations, or advanced text processing. You get access to Python libraries for machine learning, data wrangling, or NLP. For instance, you might import sklearn for classification or re for regex-based text cleansing. Python is ideal For:

  • Advanced data science, feature engineering, custom ML transformations, or unusual data-cleaning logic.

  • If you need loops, complex conditionals, or string manipulations that are easier to write in Python than SQL.

  • If you use PySpark, transformations can run in a distributed environment for very large datasets.

Tellius provides you to use Python option to:

  • Cleanse your data of invalid, missing, or inaccurate values

  • Modify your dataset according to your business goals and analysis

  • Enhance your dataset as needed with data from other datasets

Pick Python if:

  • You need advanced logic that’s awkward in SQL—like heavy string manipulation, complex conditionals, or specialized data-science libraries.

  • You’re comfortable coding in Python and want direct access to packages (e.g., Pandas, PySpark, NumPy).

  • You have iterative or row-by-row transformations that don’t translate neatly into SQL statements.

Following are some of the examples to help you get started:

def transform(dataframe):
    # use 8 spaces for indentation
       resultDataframe = dataframe.where(dataframe[‘Payment_Type’] == ‘Visa’)
       return resultDataframe
def transform(dataframe):
    # use 8 spaces for indentation
       resultDataframe = dataframe.where(dataframe[‘workclass’] == ‘Private’)
       return resultDataframe
def transform(dataframe):
    # use 8 spaces for indentation
       resultDataframe = dataframe.withColumn(‘Total’,dataframe.Qty_Sold)
       return resultDataframe

Creating and applying Python code

  1. Navigate Data → Prepare → Data.

  2. Select the required dataset and click on Edit.

  3. Above Data Pipeline, click on the Python option.

  1. To view the list of columns available in the selected dataset, click on Column List tab.

  1. Select the required Python framework: Pyspark or Pandas.

  • When working with datasets too large to fit into memory on a single machine.

  • If your data processing needs to be parallelized across multiple nodes for performance.

  • For processing cluster-based workloads stored in distributed environments (e.g., Hadoop, AWS S3, or large data warehouses).

  • Ideal for operations on terabytes/petabytes of data.

  • For small to medium data. When your dataset fits into memory on a single machine.

  • For quick, iterative data exploration and manipulation.

  • Simpler syntax and user-friendly APIs for data cleaning, transformation, and visualization.

  • Ideal for non-distributed workloads where performance isn’t a concern.

  1. To create new code, click on Create New or Write code yourself button.

  2. Alternatively, click on Generate with Kaiya button to make Tellius Kaiya generate the required for you.

  3. Once the code is ready, click on Run Validation to validate the code. When the validation is in process, the Running Validation message is displayed.

  4. Tellius validates the entered query, and if any errors are found, they will be displayed in the bottom section of the window.

  5. If the code is correct, the validation result is shown with a Successfully Validated message at the top.

  6. After clearing the errors, click on Apply to apply the code to the dataset or click on Save in Library to save to the code library in the left pane. Or, click on Cancel to discard the code window.

From v4.2, users can apply the code to the dataset without saving it to the code library first.

Editing Python code

  1. In the Python code window, search and select the required code from the already existing Code Library.

  2. Click on Edit to modify and validate the code.

  1. Click on Run Validation to validate the code. When the validation is in process, the Running Validation message is displayed.

  2. Tellius validates the entered query, and if any errors are found, they will be displayed in the bottom section of the window.

  3. If the code is correct, the validation result is shown with a Successfully Validated message at the top.

  4. Click on Apply button to apply the Python query to the dataset.

  5. Click on Update to update the code, and click on Save as New.

The following libraries have been removed and thus cannot be imported into Python during data preparation. If any of the following libraries are imported, it will result in a Validation failed error. - shlex - sh - plumbum - pexpect - fabric - envoy - commands - os - subprocess - requests

PreviousSQL TransformNextStandard Aggregation

Last updated 5 months ago

Was this helpful?

🔢
🐣
Data → Prepare → Data → Edit
Python window
Editing already existing Python code