Software Mind Logo

Software Mind

[GFA] AWS Senior Data Engineer

Posted 8 Hours Ago
Be an Early Applicant
In-Office or Remote
Hiring Remotely in Kraków, Małopolskie
Senior level
In-Office or Remote
Hiring Remotely in Kraków, Małopolskie
Senior level
Design and build metadata-driven, reusable ETL pipelines on AWS using Python/PySpark and Spark. Automate CI/CD, monitoring, and cloud provisioning; integrate streaming and batch sources; ensure data quality, performance, and platform documentation.
The summary above was generated by AI
Company Description

Software Mind develops solutions that make an impact for companies around the globe. Tech giants & unicorns, transformative projects, emerging technologies and limitless opportunities – these are a few words that describe an average day for us. Building cross-functional engineering teams that take ownership and crave more means we’re always on the lookout for talented people who bring passion and creativity to every project. Our culture embraces openness, acts with respect, shows grit & guts and combines employment with enjoyment.

Job Description

 

Project – the aim you’ll have

Our customer provides innovative solutions and insights that enable our clients to manage risk and hire the best talent. Their advanced global technology platform supports fully scalable, configurable screening programs that meet the unique needs of over 33,000 clients worldwide. Headquartered in Atlanta, GA, they have an internationally distributed workforce spanning 19 countries with about 5,500 employees. Our partner perform over 93 million screens annually in over 200 countries and territories.

We are seeking a Senior Data Engineer with strong Python/PySpark skills to join the Data Engineering team and help build our Data Analytics Platform in AWS.

Position – how you’ll contribute

  • Develop reusable, metadata-driven data pipelines
  • Automate and optimize data platform processes
  • Build integrations with data sources and data consumers
  • Add data transformation methods to shared ETL libraries
  • Write unit tests and perform code reviews to ensure code quality
  • Develop solutions for data platform monitoring and alerting (e.g., CloudWatch, third‑party tools)
  • Proactively resolve performance and data quality issues in ETL processes
  • Collaborate with infrastructure engineers to provision and configure cloud resources (VPC, IAM, S3, etc.)
  • Contribute to platform documentation and runbooks
  • Propose and implement improvements to data platform architecture

Qualifications

 

Expectations – the experience you need

  • Strong database skills (SQL, data modeling, query optimization)
  • Programming: Python / PySpark, SQL
  • Proficient in building robust data pipelines using Spark (Databricks on AWS or EMR/EMR Serverless)
  • Experienced working with large and complex datasets
  • Skilled in building reusable data transformation modules organized as Python packages
  • Familiar with Delta Lake optimization techniques on S3 (partitioning, Z-ordering, compaction) or equivalent table formats (Apache Iceberg, Hudi)
  • Experienced in developing CI/CD pipelines (e.g., GitHub Actions, Jenkins, AWS CodePipeline)
  • Experienced integrating with event brokers (Kafka, Amazon Kinesis) for ingest and streaming use-cases
  • Understanding of basic networking and security in cloud environments (VPC, subnets, security groups, IAM)
  • Familiar with Agile software development methodologies (Scrum)

 

Additional skills – the edge you have

  • Understanding of stream processing and Spark Structured Streaming or Kinesis Data Analytics
  • Experience with Infrastructure as Code (Terraform, AWS CloudFormation)
  • Experience running containerized workloads (ECS/Fargate, EKS/Kubernetes)
  • Experience building event-sourcing or CDC solutions; familiarity with Debezium a plus
  • Knowledge of AWS-native data services (AWS Glue, AWS Lambda, Amazon S3, Amazon Redshift, Amazon RDS, Amazon Athena)

Additional Information

Our offer – professional development, personal growth:

  • Flexible employment and remote work  
  • International projects with leading global clients 
  • International business trips  
  • Non-corporate atmosphere 
  • Language classes 
  • Internal & external training 
  • Private healthcare and insurance  
  • Multisport card 
  • Well-being initiatives 

Position at: Software Mind Poland

This role requires candidates to be based in Poland.

Top Skills

Python,Pyspark,Sql,Spark,Databricks,Aws Emr,Aws Emr Serverless,Delta Lake,Apache Iceberg,Hudi,Github Actions,Jenkins,Aws Codepipeline,Kafka,Amazon Kinesis,Cloudwatch,Vpc,Iam,Amazon S3,Terraform,Aws Cloudformation,Ecs,Fargate,Eks,Kubernetes,Debezium,Aws Glue,Aws Lambda,Amazon Redshift,Amazon Rds,Amazon Athena,Spark Structured Streaming,Kinesis Data Analytics

Similar Jobs

13 Hours Ago
Remote
Poland
Internship
Internship
Artificial Intelligence • Cloud • Consumer Web • Productivity • Software • App development • Data Privacy
Join Dropbox as a Software Engineering Intern to design, develop, and ship a high-priority project, collaborate with cross-functional teams, build AI proficiency (e.g., Copilot), and participate in rapid prototyping, user-focused development, and professional growth opportunities over a 12-week summer term.
Top Skills: Python,Go,Javascript,Copilot
19 Hours Ago
Remote or Hybrid
Poland
Mid level
Mid level
HR Tech • Information Technology • Professional Services • Sales • Software
As an Account Executive, you will manage the sales cycle, drive new business opportunities, and close deals in a SaaS environment targeting mid-market companies. You'll work cross-functionally, perform product demonstrations, and maintain relationships with prospects and key decision-makers.
Top Skills: Salesforce CRM
19 Hours Ago
Easy Apply
Remote or Hybrid
Kraków, Małopolskie, POL
Easy Apply
Mid level
Mid level
AdTech • Enterprise Web • Information Technology • Machine Learning • Marketing Tech • Sales
The BI Developer will analyze data, generate insights, document business needs, ensure consistent metrics, and support data-driven decision-making across teams.
Top Skills: BigQueryGCPGitLookerPower BIPythonSQLTableau

What you need to know about the Toronto Tech Scene

Although home to some of the biggest names in tech, including Google, Microsoft and Amazon, Toronto has established itself as one of the largest startup ecosystems in the world. And with over 2,000 startups — more than 30 percent of the country's total startups — Toronto continues to attract new businesses. Be it helping entrepreneurs manage their finances, simplifying business operations by automating payroll or assisting pharmaceutical companies in launching new drugs, the city's tech scene is just getting started.

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account