Data Architect

The CHI Software team is not standing still. We love our job and give it one hundred percent of us! Every new project is a challenge that we face successfully. The only thing that can stop us is… Wait, it’s nothing! The number of projects is growing, and with them, our team too. And now we need a Data Architect.

Must-Have Skills
  • Data Architecture (8+ years, 3+ as architect)
  • Enterprise data architecture — conceptual, logical, physical modeling
  • Lakehouse architecture patterns — medallion architecture, data mesh principles
  • Hybrid data architecture — on-prem ↔ cloud data movement, consistency, governance
  • Data modeling — dimensional modeling, Data Vault 2.0, or similar for analytical workloads

 

  • Cloudera / Hadoop Ecosystem (Critical)
  • Cloudera Data Platform (CDP) — architecture, administration, security
  • Apache Hive / Iceberg — table formats, partitioning strategies, compaction
  • Apache Spark on Cloudera — performance tuning, resource management
  • HDFS / Ozone — storage architecture, replication, tiering
  • Apache Ranger / Knox — security, access control, audit
  • Query & Analytics Layer
  • Starburst / Trino — federated query architecture, connector management, caching strategies
  • Tableau — data source architecture, extract vs. live connections, semantic layer design
  • SQL optimization for analytical workloads — join strategies, materialized views, caching
  • Streaming & Integration
  • Apache Kafka — topic design, schema registry (Avro/Protobuf), exactly-once semantics
  • CDC (Change Data Capture) patterns — Debezium, Kafka Connect
  • ETL/ELT pipeline design — batch and streaming patterns
  • Apache Airflow — DAG design for complex data pipelines
  • Governance & Quality
  • Informatica — data catalog, data quality, metadata management
  • Data governance frameworks — data ownership, stewardship, lineage
  • ABAC implementation for data access (attribute-based access control)
  • Data quality metrics and monitoring
  • Business glossary management

 

  • Cloud Data Architecture (AWS)
  • S3 — data lake storage patterns, lifecycle policies, storage classes
  • AWS Glue — catalog, ETL jobs
  • Lake Formation — fine-grained access control
  • Hybrid integration — AWS Direct Connect, data replication strategies
Nice-to-Have
  • Telco data models (CDR, network events, subscriber data, billing)
  • Experience migrating from traditional Hadoop to modern lakehouse
  • Data Mesh implementation experience
  • Cost optimization for large-scale data platforms
  • DAMA-DMBOK or equivalent data management certifications
Engagement Model
  • Full-time staff augmentation (embedded in team)
  • Remote acceptable
  • Must overlap with Baku working hours (GMT+4)
  • English fluency required

Our perks

  • calendar
    Covered vacation period: 20 business days and 5 days off
  • English
    Free English classes
  • clock
    Flexible working schedule
  • smile
    Truly friendly and supporting atmosphere
  • home
    Working remotely or in one of our offices
  • user
    Medical insurance for employees from Ukraine
  • legal
    Legal support

Your dream job awaits you

Apply now!

    Successfully applied!