(Senior) Data & Analytics Engineer für Berlin gesucht
Arbeits- und Stellenangebot im Regiobizz Arbeitsmarkt
Job Kategorie: IT/IT-Dienstleistung Informationstechnologie (IT)
Stellenangebot Basisdaten
- Arbeitsort:
-
DE 10115 Berlin
- Umkreis:
-
keine Angabe.
- Art der Arbeitsstelle:
-
- Letze Aktualisierung:
-
11.06.20252025-06-11
Stellenausschreibung: (Senior) Data & Analytics Engineer
- Arbeitgeber bzw.
Arbeitsvermittler
-
Gemma Analytics in Hamburg
- Branche
-
IT/IT-Dienstleistung
- Kategorie
-
Informationstechnologie (IT)
- Stellenbeschreibung
- Intro We are Gemma Analytics: a Berlin-based company specializing in
generating insights in high-performance data infrastructure. Gemma was
founded in early 2020 by two data enthusiasts. Ever since, we have
helped over 70 companies to become more data-driven and successful. We
have a fun, honest, and inclusive work environment. We are always
looking for data-minded people we can learn from. Tasks Gemma
Analytics is data-driven and helps clients to become more data-driven.
As our Senior Data & Analytics Engineers, you play a critical role in
helping our clients unlock business value from their data. You’re
not just technically strong — you’re a Data Magician who uncovers
structure in chaos and turns raw data into meaningful, actionable
insight. You dig into complex datasets, spot what others overlook, and
guide clients toward pragmatic, high-impact solutions. But your impact
doesn’t stop at client work. As a senior team member, you act as a
sparring partner and coach to your colleagues. You’re someone others
turn to for advice on technical challenges, project structure, and
best practices — and you’re excited to help them grow. You have
the opportunity to work on difficult problems while helping startups
and SMEs to make well-informed decisions based on data.Challenges: As
we are tooling-agnostic, you will touch on multiple technologies and
understand the in’s & out’s of what is currently possible in the
data landscape Collaborate with domain experts and client stakeholders
to solve data challenges across a variety of industries Support and
mentor other team members through code reviews, pair programming, and
knowledge sharing Lead internal sparring sessions and contribute to
developing team-wide best practices and scalable project structures
Technologies you’ll use Working with multiple clients, we are in
touch with many technologies, which is truly exciting. We use
state-of-the-art technologies while being fully pragmatic (we do not
crack a walnut with a sledgehammer). We follow an ELT philosophy and
divide the tasks between Data Engineering and Analytics Engineering
accordingly. The following technologies constitute our preferred data
tech stack: Data Loading For our clients, we either use a scheduler
(e.g. Apache Airflow or Prefect) and run Python DAGs with it - we also
like to work with dlt as a framework For standard connectors, we work
with Fivetran or Airbyte Cloud preferably Data Warehousing For smaller
data loads, we mostly use PostgreSQL databases For larger datasets, we
mostly work with Snowflake or BigQuery Data Transformation We love to
use dbt (data build tool) since 2018 - we can also work without it,
yet we are fans to be honest It is important to us that we work
version-controlled, peer-reviewed, with data testing, and other
engineering best practices Data Visualization For smaller businesses
with For specified needs and a centralized BI, we recommend PowerBI or
Tableau For a decentralized, self-service BI with more than 50 users,
we recommend Looker, Holistics, or ThoughtSpot We are always on the
lookout for new tools, at the moment we are excited about Lightdash,
Omni, dlt, and other tools Requirements We believe in a good mixture
of experience and upside in our team. We are looking for both types of
people equally - for this role, we require more expertise and proof of
trajectory. Besides that, we are looking for the following : 3–4
years of hands-on experience in data engineering or analytics
engineering, with a strong focus on building and maintaining robust
data pipelines and analytics-ready data models Proficient in SQL and
experienced with relational databases, capable of translating complex
business logic into clear, maintainable queries Hands-on experience
using dbt (preferably dbt Cloud) in production environments, following
best practices for modular, testable, and documented code Solid
understanding of data modeling techniques (e.g., Kimball dimensional
modeling, Data Vault, star/snowflake schema) and data warehousing
principles Experience working with modern data stack tools, such as
Snowflake, BigQuery, Airflow, Airbyte/Fivetran, Git, and CI/CD
workflows Proficient in Python (or a similar scripting language) for
use cases such as API integration, data loading, and automation Strong
communication skills in English (written and spoken), with the ability
to explain technical decisions and collaborate with both technical and
non-technical stakeholders Comfortable working in client-facing
projects, navigating ambiguity, and delivering high-quality results
with minimal oversight Experience coaching or mentoring junior team
members through code reviews, sparring, and knowledge sharing Bonus:
Familiarity with data visualization tools (e.g., Tableau, Power BI,
Looker) to support end-to-end workflows or assist analysts Bonus:
Fluency in German Benefits We are located in Berlin, close to
Nordbahnhof. We are currently 20 colleagues and will grow to 22
colleagues this year. Other perks include: We are a hybrid company
that meets in the office twice a week - one common office day and one
flexible day We allow for intra-EU workcations for up to 3 months a
year (extra-EU workcations also if this is allowed) We have an honest,
inclusive work environment and want to nurture this environment We
don’t compromise on equipment - a powerful Laptop, extra screens,
and all the tools you need to be effective We will surround you with
great people who love to solve (mostly data) riddles We believe in
efficient working hours rather than long working hours - we focus on
the output rather than the input We learn and share during meetups,
lunch & learn sessions and are open to further initiatives We pay a
market-friendly salary and we additionally distribute at least 20% of
profits to our employees We are fast-growing and have technology at
our core, yet we do not rely on a VC and operate profitably We have a
great yearly offsite event that brings us all together for a full
week, enjoying good food, and having a good time (2021: Austria, 2022:
Czech Republic, 2023+2024: Germany, 2025: Spain) Closing How you’ll
get here CV Screening Phone/Coffee/Tea Initial Conversation Hiring
Test @home Interviews with 2-3 future colleagues Reference calls Offer
+ Hired Looking forward to your application :)
- Qualifikation
- Arbeitskräfte
- Verdienst:
- n.a.
- Bewerbung an
- Gemma Analytics
Am Strandkai 1
De 20457 Hamburg
Stellenangebot powered by

Die Veröffentlichung dieses Stellenangebotes bei regiobizz.de erfolgt mit freundlicher Genehmigung von GermanPersonnel im Namen des Stellenanbieters. Eine gewerbliche Nutzung dieser Daten sowie deren Veröffentlichung in jeder Form ist ohne ausdrückliche Genehmigung von GermanPersonnel strengstens untersagt.