Python

Python

115 posts
Manage Settings and Environment Variables in FastAPI
Academy Membership FastAPIPython

Manage Settings and Environment Variables in FastAPI

Introduction FastAPI is a modern, fast (high-performance) web framework for building APIs with Python. When building real-world applications, you’ll often need to manage sensitive information like API keys, database URLs, or feature toggles that should vary across environments (development, testing, production). Using environment variables and Pydantic’s BaseSettings is...

PySpark - Get statistical Properties of a DataFrame
Academy Membership PySparkPython

PySpark - Get statistical Properties of a DataFrame

Introduction When working with PySpark DataFrames, understanding the statistical properties of your data is crucial for data exploration and preprocessing. PySpark provides the describe() and summary() functions to generate useful summary statistics. In this tutorial, we’ll explore how to use both functions to get insights into our dataset. 📥 Import...

Install and use dlt (data load tool)
Academy Membership dltData Engineering

Install and use dlt (data load tool)

Introduction dlt (data load tool) is a powerful Python package that simplifies data ingestion and helps you build efficient data pipelines. In the Extract, Load, Transform (ELT) process, dlt is particularly suited for the Extract (E) and Load (L) stages. In this tutorial, we'll guide you through the...

PySpark - Convert Column Data Types of a DataFrame
Academy Membership PySparkDP-600

PySpark - Convert Column Data Types of a DataFrame

Introduction When working with PySpark DataFrames, handling different data types correctly is essential for data preprocessing. Mismatched or incorrect data types can lead to errors in Spark operations such as filtering, aggregations, and machine learning workflows. In this tutorial, we’ll explore how to convert column data types in a...

PySpark - Replace Empty Strings with Null Values
Academy Membership PySparkPython

PySpark - Replace Empty Strings with Null Values

Introduction When working with PySpark DataFrames, handling missing or empty values is a common task in data preprocessing. In many cases, empty strings ("") should be treated as null values for better compatibility with Spark operations, such as filtering, aggregations, and machine learning workflows. In this tutorial, we’ll...

You’ve successfully subscribed to Deep Learning Nerds | The ultimate Learning Platform for AI and Data Science
Welcome back! You’ve successfully signed in.
Great! You’ve successfully signed up.
Success! Your email is updated.
Your link has expired
Success! Check your email for magic link to sign-in.