Technical Architect-Databricks

Publication Date:  Oct 4, 2025
Ref. No:  531371
Location: 

Pune, IN

About Atos Group

 

Atos Group is a global leader in digital transformation with c. 70,000 employees and annual revenue of c. € 10 billion, operating in 67 countries under two brands — Atos for services and Eviden for products. European number one in cybersecurity, cloud and high-performance computing, Atos Group is committed to a secure and decarbonized future and provides tailored AI-powered, end-to-end solutions for all industries. Atos is a SE (Societas Europaea) and listed on Euronext Paris.

 

The purpose of Atos is to help design the future of the information space. Its expertise and services support the development of knowledge, education and research in a multicultural approach and contribute to the development of scientific and technological excellence. Across the world, the Group enables its customers and employees, and members of societies at large to live, work and develop sustainably, in a safe and secure information space.

Role Overview:

The Technical Architect - Databricks designs and implements scalable data architectures and solutions. The jobholder has expertise in Databricks Lakehouse, data modeling, and cloud integration, ensuring high performance, security, and reliability.

Responsibilities:

  • Design and implement Databricks-based data architectures to meet business requirements.
  • Develop and optimize data pipelines using PySpark, Scala, or SQL.
  • Establish the Databricks Lakehouse architecture for batch and streaming data.
  • Collaborate with cross-functional teams to integrate Databricks with cloud platforms (e.g., AWS, Azure, GCP).
  • Ensure data security and compliance with best practices.
  • Monitor and troubleshoot Databricks environments for performance and reliability.
  • Stay updated on Databricks advancements and industry trends.

Key Technical Skills & Responsibilities

  • 12+  years of experience in data engineering using Databricks or Apache Spark-based platforms.
  • Proven track record of building and optimizing ETL/ELT pipelines for batch and streaming data ingestion.
  • Hands-on experience with Azure services such as Azure Data Factory, Azure Data Lake Storage, Azure Synapse Analytics, or Azure SQL Data Warehouse.
  • Proficiency in programming languages such as Python, Scala, or SQL for data processing and transformation.
  • Expertise in Spark (PySpark, Spark SQL, or Scala) and Databricks notebooks for large-scale data processing.
  • Familiarity with Delta Lake, Delta Live Tables, and medallion architecture for data lakehouse implementations.
  • Build and query deltalake storage solutions 
  • Process streaming data with Azure Databricks structured streaming
  • Design Azure Databricks security and data protection solutions
  • Flatten nested structures and explode arrays with spark
  • Transfer data outside using sparkpools using pyspark connector
  • Optimizing spark jobs
  • Implementing best practices in spark/databricks
  • Experience with orchestration tools like Azure Data Factory or Databricks Jobs for scheduling and automation.
  • Knowledge of Git for source control and CI/CD integration for Databricks workflows, cost optimization, performance tuning.
  • Familiarity with Unity Catalog, RBAC, or enterprise-level Databricks setups.
  • Ability to create reusable components, templates, and documentation to standardize data engineering workflows.
  • Solutioning and presales - Architecting frameworks, defining roadmaps, and engaging with stakeholders.
  • Experience in defining data strategy, evaluating new tools/technologies, and driving adoption across the organization.
  • Must have experience of working with streaming data sources and Kafka (preferred).

Eligibility Criteria:

  • Bachelor’s degree in computer science, Information Technology, or related field
  • Proven experience as a Databricks Architect or similar role
  • Complete knowledge in Azure Databricks platform architecture
  • Databricks certification (e.g., Certified Data Engineer, Associate Developer)
  • Expertise in Python/Scala/ SQL/R
  • Experience with cloud platforms like AWS, Azure, or GCP
  • Strong understanding of data modeling and cloud integration
  • Experience with cluster sizing and security implementation
  • Excellent problem-solving and communication skills