About the Company
Boycor is a technology consulting firm specializing in delivering data-driven solutions for the financial sector. The company works closely with clients on projects that require high expertise in data engineering, cloud technologies, and advanced analytics, with a focus on Anti-Money Laundering (AML) and compliance processes.
Boycor fosters a collaborative and flexible work environment where technical autonomy, innovation, and continuous professional growth are highly valued.
About the Role
The Data Engineer will join a specialized AML team, designing, developing, and optimizing distributed data pipelines to process large volumes of financial data. The role involves working with Spark (Scala), Python, Airflow, and AWS S3, as well as integrating data into Elasticsearch for analysis and reporting.
This position is 100% remote with optional hybrid work in Málaga or Logroño, offering flexibility and long-term engagement.
Responsibilities
- Design and develop Spark (Scala) pipelines for high-volume financial data, applying advanced transformations and performance optimizations
- Build and maintain DAGs in Airflow, ensuring proper orchestration, dependency management, and monitoring of critical processes
- Develop Python utilities to support Airflow orchestration and automate repetitive tasks
- Manage data ingestion and structuring in AWS S3, ensuring data integrity, efficiency, and accessibility
- Integrate data into Elasticsearch for use in Kibana and antifraud analysis
- Participate in migration of processes from Cloudera (Spark 2.4) to Databricks (Spark 3.x), optimizing code and applying performance best practices
- Collaborate with clients to manage ongoing developments and incidents, providing technical autonomy and proposing solutions
Requirements
- Strong experience developing and optimizing distributed processes in Spark with Scala, including partitioning, caching, and performance tuning
- Proficiency in Python for scripting, automation, and orchestration support
- Experience with cloud environments and S3 storage, including data ingestion, structuring, and efficient access patterns
- Knowledge of version control and collaborative development using Git, along with CI/CD best practices
- Proactive problem-solving mindset with ability to communicate effectively with clients
- Experience with Airflow for complex pipeline orchestration is highly valued
- Bonus: Experience with Databricks, scheduling tools like Control-M, or projects in the banking/financial sector
Benefits
- Permanent contract on a stable, long-term project
- 100% remote work, with optional hybrid presence in Málaga or Logroño
- Flexible working hours: Monday–Thursday 8.5h, Friday 6h; intensive 7h schedule during July and August
