Lead AI Consultant Location: Bristol or Swindon (2 days a week on-site) Salary: £70,000 - £130,000 (based on experience) Sanderson are seeking an experienced Lead AI Consultant to join a leading SaaS platform provider. This is a unique opportunity to be part of a dynamic team and play a pivotal role in shaping the strategic direction of AI implementation within the business. About the Role: As the Lead AI Consultant, you will report directly to the CTO and be responsible for driving AI adoption, developing "big data" solutions, and defining API strategies. You will collaborate with senior product, technical, and operational managers to enhance the product suite with cutting-edge AI technologies. Key Responsibilities: Develop and execute the technical integration strategy. Create comprehensive "big data" solutions across products. Define internal and external API strategies. Enhance existing products through AI technologies. Design, develop, and implement AI solutions. Integrate AI with client data, APIs, and production systems. Conduct data analysis and build machine learning models. Utilize modern ML and AI libraries and frameworks. Deliver reliable solutions collaboratively. Desired Experience: Significant experience as an AI Consultant, Machine Learning Engineer, or similar role. Proven track record in start-up or dynamic environments. Hands-on experience with cloud-based AI and ML solutions (eg Azure). Knowledge of generative AI applications and embedding models. Proficiency in MLOps and DevOps practices. Strong programming skills in Python and C#. Experience with Big Data technologies and ETL processes. Familiarity with distributed computing frameworks like Kubernetes. Why Join: Drive AI innovation at a leading SaaS provider. Work with a dynamic, forward-thinking team. Engage in challenging projects that impact business growth. Enjoy a high degree of autonomy in defining AI strategy.
17/05/2024
Full time
Lead AI Consultant Location: Bristol or Swindon (2 days a week on-site) Salary: £70,000 - £130,000 (based on experience) Sanderson are seeking an experienced Lead AI Consultant to join a leading SaaS platform provider. This is a unique opportunity to be part of a dynamic team and play a pivotal role in shaping the strategic direction of AI implementation within the business. About the Role: As the Lead AI Consultant, you will report directly to the CTO and be responsible for driving AI adoption, developing "big data" solutions, and defining API strategies. You will collaborate with senior product, technical, and operational managers to enhance the product suite with cutting-edge AI technologies. Key Responsibilities: Develop and execute the technical integration strategy. Create comprehensive "big data" solutions across products. Define internal and external API strategies. Enhance existing products through AI technologies. Design, develop, and implement AI solutions. Integrate AI with client data, APIs, and production systems. Conduct data analysis and build machine learning models. Utilize modern ML and AI libraries and frameworks. Deliver reliable solutions collaboratively. Desired Experience: Significant experience as an AI Consultant, Machine Learning Engineer, or similar role. Proven track record in start-up or dynamic environments. Hands-on experience with cloud-based AI and ML solutions (eg Azure). Knowledge of generative AI applications and embedding models. Proficiency in MLOps and DevOps practices. Strong programming skills in Python and C#. Experience with Big Data technologies and ETL processes. Familiarity with distributed computing frameworks like Kubernetes. Why Join: Drive AI innovation at a leading SaaS provider. Work with a dynamic, forward-thinking team. Engage in challenging projects that impact business growth. Enjoy a high degree of autonomy in defining AI strategy.
Request Technology - Craig Johnson
Chicago, Illinois
*We are unable to sponsor for this permanent Full time role* *Position is bonus eligible* Prestigious Financial Institution is currently seeking a Principal Java Risk Management Software Engineer. Candidate will develop and maintain risk models for margin, clearing fund and stress testing with the focus on developing and maintaining risk model software in production, and environments and infrastructure used in model implementation and testing. Candidate will collaborate with other developers, quantitative analysts, business users, data & technology staff to expand the technical capabilities for model development, back-testing and monitoring. Responsibilities: Develop and maintain software and environments used to implement and test systems for pricing, margin risk and stress testing of financial products and derivatives. Configure and manage resources in the local and AWS cloud environments and deploy QRM's software on these resources. Develop CI/CD pipelines. Configure, execute, and monitor execution pipelines for model testing, back-testing and monitoring. Contribute to development of QRM's databases and ETLs. Integrate model prototypes, model library and model testing tools using best industry practices and innovations. Create unit and integration tests; build and enhance test automation tools. Participate in code reviews and demo accomplishments. Write technical documentation and user manuals. Provide production support and perform troubleshooting. Qualifications: Strong programming skills. Able to read and/or write code using a programming language (eg, Java, C++, Python, etc.) in a collaborative software development setting: The role requires advanced coding, database and environment manipulation skills. Track record of complex production implementations and a demonstrated ability in developing and maintaining enterprise level software, including in the cloud environment. Proficiency in technical and/or scientific documentation (eg, white papers, user guides, etc.) Strong problem-solving skills: Be able to accurately identify a problem's source, severity, and impact to determine possible solutions and needed resources. Experience with Agile/SCRUM or another rapid development framework. Financial products knowledge is a plus: understanding of markets and financial derivatives in equities, interest rate, and commodity products. Background in Financial mathematics is a plus: derivatives pricing models, stochastic calculus, statistics and probability theory, linear algebra. Master's degree or equivalent in a computational or numerical field such as computer science, information systems, mathematics, physics 10+ years of experience as a software developer with exposure to the cloud or high-performance computing areas Technical Skills: Proficiency in Java (preferred) or another object-oriented language is required, including effective application of design patterns and best coding practices. DevOps experience, with a good command of CI/CD process and tools (eg, Git, GitHub, Gradle, Jenkins, Docker, Helm, Harness). Experience in containerized deployment in cloud environments. Experienced with cloud technology (AWS preferred), infrastructure-as-code (eg Terraform), managing and orchestrating containerized workloads (eg Kubernetes). Experience with logging, profiling, monitoring, telemetry (eg Splunk, OpenTelemetry). Good command of database technology and query languages (SQL) and non-relational DB and other Big Data technology, including efficient storage and serialization protocols (eg Parquet, Avro, Protocol Buffers). Experience with automated quality assurance frameworks (eg, Junit, TestNG, PyTest, etc.). Experience with high performance and distributed computing. Experience with productivity tools such as Jira, Confluence, MS Office. Experience with Scripting languages such as Python is a plus. Experience with numerical libraries and/or scientific computing is a plus.
16/05/2024
Full time
*We are unable to sponsor for this permanent Full time role* *Position is bonus eligible* Prestigious Financial Institution is currently seeking a Principal Java Risk Management Software Engineer. Candidate will develop and maintain risk models for margin, clearing fund and stress testing with the focus on developing and maintaining risk model software in production, and environments and infrastructure used in model implementation and testing. Candidate will collaborate with other developers, quantitative analysts, business users, data & technology staff to expand the technical capabilities for model development, back-testing and monitoring. Responsibilities: Develop and maintain software and environments used to implement and test systems for pricing, margin risk and stress testing of financial products and derivatives. Configure and manage resources in the local and AWS cloud environments and deploy QRM's software on these resources. Develop CI/CD pipelines. Configure, execute, and monitor execution pipelines for model testing, back-testing and monitoring. Contribute to development of QRM's databases and ETLs. Integrate model prototypes, model library and model testing tools using best industry practices and innovations. Create unit and integration tests; build and enhance test automation tools. Participate in code reviews and demo accomplishments. Write technical documentation and user manuals. Provide production support and perform troubleshooting. Qualifications: Strong programming skills. Able to read and/or write code using a programming language (eg, Java, C++, Python, etc.) in a collaborative software development setting: The role requires advanced coding, database and environment manipulation skills. Track record of complex production implementations and a demonstrated ability in developing and maintaining enterprise level software, including in the cloud environment. Proficiency in technical and/or scientific documentation (eg, white papers, user guides, etc.) Strong problem-solving skills: Be able to accurately identify a problem's source, severity, and impact to determine possible solutions and needed resources. Experience with Agile/SCRUM or another rapid development framework. Financial products knowledge is a plus: understanding of markets and financial derivatives in equities, interest rate, and commodity products. Background in Financial mathematics is a plus: derivatives pricing models, stochastic calculus, statistics and probability theory, linear algebra. Master's degree or equivalent in a computational or numerical field such as computer science, information systems, mathematics, physics 10+ years of experience as a software developer with exposure to the cloud or high-performance computing areas Technical Skills: Proficiency in Java (preferred) or another object-oriented language is required, including effective application of design patterns and best coding practices. DevOps experience, with a good command of CI/CD process and tools (eg, Git, GitHub, Gradle, Jenkins, Docker, Helm, Harness). Experience in containerized deployment in cloud environments. Experienced with cloud technology (AWS preferred), infrastructure-as-code (eg Terraform), managing and orchestrating containerized workloads (eg Kubernetes). Experience with logging, profiling, monitoring, telemetry (eg Splunk, OpenTelemetry). Good command of database technology and query languages (SQL) and non-relational DB and other Big Data technology, including efficient storage and serialization protocols (eg Parquet, Avro, Protocol Buffers). Experience with automated quality assurance frameworks (eg, Junit, TestNG, PyTest, etc.). Experience with high performance and distributed computing. Experience with productivity tools such as Jira, Confluence, MS Office. Experience with Scripting languages such as Python is a plus. Experience with numerical libraries and/or scientific computing is a plus.
Request Technology - Craig Johnson
Chicago, Illinois
*We are unable to sponsor for this permanent Full time role* *Position is bonus eligible* Prestigious Financial Institution is currently seeking a Director of Risk Management Software Engineering. Candidate will be responsible for functions within Quantitative Risk Management for developing and maintaining risk models for margin, clearing fund and stress testing with the focus on developing and maintaining risk model software in production, and environments and infrastructure used in model implementation and testing. Responsibilities: Collaborate with other developers, quantitative analysts, business users, data & technology staff to expand QRM's technical capabilities for model development, back-testing and monitoring. Develop and maintain software and environments used to implement and test systems for pricing, margin risk and stress testing of financial products and derivatives. Configure and manage resources in the local and AWS cloud environments and deploy QRM's software on these resources. Develop CI/CD pipelines. Configure, execute, and monitor execution pipelines for model testing, back-testing and monitoring. Contribute to development of QRM's databases and ETLs. Integrate model prototypes, model library and model testing tools using best industry practices and innovations. Create unit and integration tests; build and enhance test automation tools. Participate in code reviews and demo accomplishments. Write technical documentation and user manuals. Provide production support and perform troubleshooting. Provide hands-on technical leadership and active coordination of tasks and priorities. Provide guidance and support for the team and reporting for the management. Qualifications: Strong programming skills. Able to read and/or write code using a programming language (eg, Java, C++, Python, etc.) in a collaborative software development setting: The role requires advanced coding, database and environment manipulation skills. Track record of complex production implementations and a demonstrated ability in developing and maintaining enterprise level software, including in the cloud environment. Proficiency in technical and/or scientific documentation (eg, white papers, user guides, etc.) Strong problem-solving skills: Be able to accurately identify a problem's source, severity, and impact to determine possible solutions and needed resources. Experience with Agile/SCRUM or another rapid development framework. Financial products knowledge is a plus: understanding of markets and financial derivatives in equities, interest rate, and commodity products. Background in Financial mathematics is a plus: derivatives pricing models, stochastic calculus, statistics and probability theory, linear algebra. Master's degree or equivalent in a computational or numerical field such as computer science, information systems, mathematics, physics 10+ years of experience as a software developer with exposure to the cloud or high-performance computing areas Technical Skills: Proficiency in Java (preferred) or another object-oriented language is required, including effective application of design patterns and best coding practices. DevOps experience, with a good command of CI/CD process and tools (eg, Git, GitHub, Gradle, Jenkins, Docker, Helm, Harness). Experience in containerized deployment in cloud environments. Experienced with cloud technology (AWS preferred), infrastructure-as-code (eg Terraform), managing and orchestrating containerized workloads (eg Kubernetes). Experience with logging, profiling, monitoring, telemetry (eg Splunk, OpenTelemetry). Good command of database technology and query languages (SQL) and non-relational DB and other Big Data technology, including efficient storage and serialization protocols (eg Parquet, Avro, Protocol Buffers). Experience with automated quality assurance frameworks (eg, Junit, TestNG, PyTest, etc.). Experience with high performance and distributed computing. Experience with productivity tools such as Jira, Confluence, MS Office. Experience with Scripting languages such as Python is a plus. Experience with numerical libraries and/or scientific computing is a plus.
16/05/2024
Full time
*We are unable to sponsor for this permanent Full time role* *Position is bonus eligible* Prestigious Financial Institution is currently seeking a Director of Risk Management Software Engineering. Candidate will be responsible for functions within Quantitative Risk Management for developing and maintaining risk models for margin, clearing fund and stress testing with the focus on developing and maintaining risk model software in production, and environments and infrastructure used in model implementation and testing. Responsibilities: Collaborate with other developers, quantitative analysts, business users, data & technology staff to expand QRM's technical capabilities for model development, back-testing and monitoring. Develop and maintain software and environments used to implement and test systems for pricing, margin risk and stress testing of financial products and derivatives. Configure and manage resources in the local and AWS cloud environments and deploy QRM's software on these resources. Develop CI/CD pipelines. Configure, execute, and monitor execution pipelines for model testing, back-testing and monitoring. Contribute to development of QRM's databases and ETLs. Integrate model prototypes, model library and model testing tools using best industry practices and innovations. Create unit and integration tests; build and enhance test automation tools. Participate in code reviews and demo accomplishments. Write technical documentation and user manuals. Provide production support and perform troubleshooting. Provide hands-on technical leadership and active coordination of tasks and priorities. Provide guidance and support for the team and reporting for the management. Qualifications: Strong programming skills. Able to read and/or write code using a programming language (eg, Java, C++, Python, etc.) in a collaborative software development setting: The role requires advanced coding, database and environment manipulation skills. Track record of complex production implementations and a demonstrated ability in developing and maintaining enterprise level software, including in the cloud environment. Proficiency in technical and/or scientific documentation (eg, white papers, user guides, etc.) Strong problem-solving skills: Be able to accurately identify a problem's source, severity, and impact to determine possible solutions and needed resources. Experience with Agile/SCRUM or another rapid development framework. Financial products knowledge is a plus: understanding of markets and financial derivatives in equities, interest rate, and commodity products. Background in Financial mathematics is a plus: derivatives pricing models, stochastic calculus, statistics and probability theory, linear algebra. Master's degree or equivalent in a computational or numerical field such as computer science, information systems, mathematics, physics 10+ years of experience as a software developer with exposure to the cloud or high-performance computing areas Technical Skills: Proficiency in Java (preferred) or another object-oriented language is required, including effective application of design patterns and best coding practices. DevOps experience, with a good command of CI/CD process and tools (eg, Git, GitHub, Gradle, Jenkins, Docker, Helm, Harness). Experience in containerized deployment in cloud environments. Experienced with cloud technology (AWS preferred), infrastructure-as-code (eg Terraform), managing and orchestrating containerized workloads (eg Kubernetes). Experience with logging, profiling, monitoring, telemetry (eg Splunk, OpenTelemetry). Good command of database technology and query languages (SQL) and non-relational DB and other Big Data technology, including efficient storage and serialization protocols (eg Parquet, Avro, Protocol Buffers). Experience with automated quality assurance frameworks (eg, Junit, TestNG, PyTest, etc.). Experience with high performance and distributed computing. Experience with productivity tools such as Jira, Confluence, MS Office. Experience with Scripting languages such as Python is a plus. Experience with numerical libraries and/or scientific computing is a plus.
Director, Software Engineering - Quantitative Risk Management Applications SALARY: $200k - $230k flex plus 27% bonus LOCATION: Chicago, il Hybrid 3 days onsite, 2 days remote You will manage six plus people and help build the framewrok within the quantitative management platform developing software applications and solutions. Java C++ python automation devops cicd aws terraform Kubernetes SQL docker helm masters or Phd This role is responsible for one or more functions within Quantitative Risk Management (QRM) who develops and maintains risk models for margin, clearing fund and stress testing with the focus on developing and maintaining risk model software in production, and environments and infrastructure used in model implementation and testing. This role will collaborate with other developers, quantitative analysts, business users, data & technology staff to expand QRM's technical capabilities for model development, backtesting and monitoring. Develop and maintain software and environments used to implement and test systems for pricing, margin risk and stress testing of financial products and derivatives. Configure and manage resources in the local and AWS cloud environments and deploy QRM's software on these resources. Develop CI/CD pipelines. Configure, execute, and monitor execution pipelines for model testing, backtesting and monitoring. Contribute to development of QRM's databases and ETLs. Integrate model prototypes, model library and model testing tools using best industry practices and innovations. Create unit and integration tests; build and enhance test automation tools. Participate in code reviews and demo accomplishments. Write technical documentation and user manuals. Provide production support and perform troubleshooting. Qualifications: Strong programming skills. Able to read and/or write code using a programming language (eg, Java, C++, Python, etc.) in a collaborative software development setting: The role requires advanced coding, database and environment manipulation skills. Track record of complex production implementations and a demonstrated ability in developing and maintaining enterprise level software, including in the cloud environment. Proficiency in technical and/or scientific documentation (eg, white papers, user guides, etc.) Strong problem-solving skills: Be able to accurately identify a problem's source, severity, and impact to determine possible solutions and needed resources. Experience with Agile/SCRUM or another rapid development framework. Financial products knowledge is a plus: understanding of markets and financial derivatives in equities, interest rate, and commodity products. Background in Financial mathematics is a plus: derivatives pricing models, stochastic calculus, statistics and probability theory, linear algebra. Technical Skills: Proficiency in Java (preferred) or another object-oriented language is required, including effective application of design patterns and best coding practices. DevOps experience, with a good command of CI/CD process and tools (eg, Git, GitHub, Gradle, Jenkins, Docker, Helm, Harness). Experience in containerized deployment in cloud environments. Experienced with cloud technology (AWS preferred), infrastructure-as-code (eg Terraform), managing and orchestrating containerized workloads (eg Kubernetes). Experience with logging, profiling, monitoring, telemetry (eg Splunk, OpenTelemetry). Good command of database technology and query languages (SQL) and non-relational DB and other Big Data technology, including efficient storage and serialization protocols (eg Parquet, Avro, Protocol Buffers). Experience with automated quality assurance frameworks (eg, Junit, TestNG, PyTest, etc.). Experience with high performance and distributed computing. Education and/or Experience: Master's degree or equivalent in a computational or numerical field such as computer science, information systems, mathematics, physics 10+ years of experience as a software developer with exposure to the cloud or high-performance computing areas
16/05/2024
Full time
Director, Software Engineering - Quantitative Risk Management Applications SALARY: $200k - $230k flex plus 27% bonus LOCATION: Chicago, il Hybrid 3 days onsite, 2 days remote You will manage six plus people and help build the framewrok within the quantitative management platform developing software applications and solutions. Java C++ python automation devops cicd aws terraform Kubernetes SQL docker helm masters or Phd This role is responsible for one or more functions within Quantitative Risk Management (QRM) who develops and maintains risk models for margin, clearing fund and stress testing with the focus on developing and maintaining risk model software in production, and environments and infrastructure used in model implementation and testing. This role will collaborate with other developers, quantitative analysts, business users, data & technology staff to expand QRM's technical capabilities for model development, backtesting and monitoring. Develop and maintain software and environments used to implement and test systems for pricing, margin risk and stress testing of financial products and derivatives. Configure and manage resources in the local and AWS cloud environments and deploy QRM's software on these resources. Develop CI/CD pipelines. Configure, execute, and monitor execution pipelines for model testing, backtesting and monitoring. Contribute to development of QRM's databases and ETLs. Integrate model prototypes, model library and model testing tools using best industry practices and innovations. Create unit and integration tests; build and enhance test automation tools. Participate in code reviews and demo accomplishments. Write technical documentation and user manuals. Provide production support and perform troubleshooting. Qualifications: Strong programming skills. Able to read and/or write code using a programming language (eg, Java, C++, Python, etc.) in a collaborative software development setting: The role requires advanced coding, database and environment manipulation skills. Track record of complex production implementations and a demonstrated ability in developing and maintaining enterprise level software, including in the cloud environment. Proficiency in technical and/or scientific documentation (eg, white papers, user guides, etc.) Strong problem-solving skills: Be able to accurately identify a problem's source, severity, and impact to determine possible solutions and needed resources. Experience with Agile/SCRUM or another rapid development framework. Financial products knowledge is a plus: understanding of markets and financial derivatives in equities, interest rate, and commodity products. Background in Financial mathematics is a plus: derivatives pricing models, stochastic calculus, statistics and probability theory, linear algebra. Technical Skills: Proficiency in Java (preferred) or another object-oriented language is required, including effective application of design patterns and best coding practices. DevOps experience, with a good command of CI/CD process and tools (eg, Git, GitHub, Gradle, Jenkins, Docker, Helm, Harness). Experience in containerized deployment in cloud environments. Experienced with cloud technology (AWS preferred), infrastructure-as-code (eg Terraform), managing and orchestrating containerized workloads (eg Kubernetes). Experience with logging, profiling, monitoring, telemetry (eg Splunk, OpenTelemetry). Good command of database technology and query languages (SQL) and non-relational DB and other Big Data technology, including efficient storage and serialization protocols (eg Parquet, Avro, Protocol Buffers). Experience with automated quality assurance frameworks (eg, Junit, TestNG, PyTest, etc.). Experience with high performance and distributed computing. Education and/or Experience: Master's degree or equivalent in a computational or numerical field such as computer science, information systems, mathematics, physics 10+ years of experience as a software developer with exposure to the cloud or high-performance computing areas
NO SPONSORSHIP Company is located in Chicago - Can be hybrid or remote if needed Contract to Hire role Data Quality Test Engineer This can be a straight contract role or a contract to hire role. The Test Engineer will be responsible for ensuring the successful implementation of new features as well as improvement of existing functionality for data architecture projects. This includes Integration testing, Customer Data Quality testing, Data verification testing, Regression testing etc. This is a hands-on testing position which requires deep knowledge and understanding of all aspects of developing and executing Back End testing. The Test Engineer must be an advocate for all things Testing and Quality Assurance while driving and implementing best practices. The ideal candidate must possess strong analytical and critical thinking abilities. Data testing skills are required. Position Requirements Bachelor's degree At least 3-5 years of testing experience in data Strong experience with writing queries in relational databases like Oracle and SQL server. Familiar with SQL server queries including calling stored procedures by passing parameters, functions, views, and indexes to be used during ETL process. Possess extensive and in-depth knowledge of developing SQL scripts using SQL functions, grouping operations, sub queries, analytical function and joins to test ETL projects. ETL testing using SQL server Integration Services (SSIS) and SSRS. Experience in testing database applications to validate source to destination data movement and transformation. Knowledge of automated data testing is preferred. Hands-on experience in Planning and Manual test execution. Experience in Regression Test Strategies, Regression suite maintenance and execution. Hands-on experience with Jira, ALM, TFS/ADO or similar systems of record. Understanding of Development and Test cycles, including respective best practices. Hands-on experience in Agile and Waterfall methodologies. Ability to troubleshoot issues, identifying root causes and support development team in development of resolutions. Strong communication and Client/Business interfacing and interpersonal skills are a must.
16/05/2024
NO SPONSORSHIP Company is located in Chicago - Can be hybrid or remote if needed Contract to Hire role Data Quality Test Engineer This can be a straight contract role or a contract to hire role. The Test Engineer will be responsible for ensuring the successful implementation of new features as well as improvement of existing functionality for data architecture projects. This includes Integration testing, Customer Data Quality testing, Data verification testing, Regression testing etc. This is a hands-on testing position which requires deep knowledge and understanding of all aspects of developing and executing Back End testing. The Test Engineer must be an advocate for all things Testing and Quality Assurance while driving and implementing best practices. The ideal candidate must possess strong analytical and critical thinking abilities. Data testing skills are required. Position Requirements Bachelor's degree At least 3-5 years of testing experience in data Strong experience with writing queries in relational databases like Oracle and SQL server. Familiar with SQL server queries including calling stored procedures by passing parameters, functions, views, and indexes to be used during ETL process. Possess extensive and in-depth knowledge of developing SQL scripts using SQL functions, grouping operations, sub queries, analytical function and joins to test ETL projects. ETL testing using SQL server Integration Services (SSIS) and SSRS. Experience in testing database applications to validate source to destination data movement and transformation. Knowledge of automated data testing is preferred. Hands-on experience in Planning and Manual test execution. Experience in Regression Test Strategies, Regression suite maintenance and execution. Hands-on experience with Jira, ALM, TFS/ADO or similar systems of record. Understanding of Development and Test cycles, including respective best practices. Hands-on experience in Agile and Waterfall methodologies. Ability to troubleshoot issues, identifying root causes and support development team in development of resolutions. Strong communication and Client/Business interfacing and interpersonal skills are a must.
Associate Principal, Software Programming - Quantitative Risk Management Area - Associate Principal, Software Engineering - Automating Risk Models On site 3 days a week Salary - $185 - $195K + Bonus Looking for a hard core developer who works within the quantitative risk management and cab develop applications and solutions for the QRM team. You will not build models, you will automate models You will need to come from a financial institute, trading company, exchange, etc. Develop hardcore applications You will need to have CICD pipelines, Infrastructure as a Code, Kubernetes, Terraform, etc. Preferably having Java, Python, C++ Configure and manage resources in the local and AWS cloud environments and deploy QRM's software on these resources. Develop CI/CD pipelines. Contribute to development of QRM's databases and ETLs. Integrate model prototypes, model library and model testing tools using best industry practices and innovations. Create unit and integration tests; build and enhance test automation tools. Participate in code reviews and demo accomplishments. Write technical documentation and user manuals. Provide production support and perform troubleshooting. Strong programming skills. Able to read and/or write code using a programming language (eg, Java, C++, Python, etc.) in a collaborative software development setting: The role requires advanced coding, database and environment manipulation skills. cloud environment. Financial products knowledge is a plus: understanding of markets and financial derivatives in equities, interest rate, and commodity products. Background in Financial mathematics is a plus: derivatives pricing models, stochastic calculus, statistics and probability theory, linear algebra. Technical Skills: Proficiency in Java (preferred) or another object-oriented language is required, including effective application of design patterns and best coding practices. DevOps experience, with a good command of CI/CD process and tools (eg, Git, GitHub, Gradle, Jenkins, Docker, Helm, Harness). Experience in containerized deployment in cloud environments. Experienced with cloud technology (AWS preferred), infrastructure-as-code (eg Terraform), managing and orchestrating containerized workloads (eg Kubernetes). Education and/or Experience: Master's degree or equivalent in a computational or numerical field such as computer science, information systems, mathematics, physics 7+ years of experience as a software developer with exposure to the cloud or high-performance computing areas
16/05/2024
Full time
Associate Principal, Software Programming - Quantitative Risk Management Area - Associate Principal, Software Engineering - Automating Risk Models On site 3 days a week Salary - $185 - $195K + Bonus Looking for a hard core developer who works within the quantitative risk management and cab develop applications and solutions for the QRM team. You will not build models, you will automate models You will need to come from a financial institute, trading company, exchange, etc. Develop hardcore applications You will need to have CICD pipelines, Infrastructure as a Code, Kubernetes, Terraform, etc. Preferably having Java, Python, C++ Configure and manage resources in the local and AWS cloud environments and deploy QRM's software on these resources. Develop CI/CD pipelines. Contribute to development of QRM's databases and ETLs. Integrate model prototypes, model library and model testing tools using best industry practices and innovations. Create unit and integration tests; build and enhance test automation tools. Participate in code reviews and demo accomplishments. Write technical documentation and user manuals. Provide production support and perform troubleshooting. Strong programming skills. Able to read and/or write code using a programming language (eg, Java, C++, Python, etc.) in a collaborative software development setting: The role requires advanced coding, database and environment manipulation skills. cloud environment. Financial products knowledge is a plus: understanding of markets and financial derivatives in equities, interest rate, and commodity products. Background in Financial mathematics is a plus: derivatives pricing models, stochastic calculus, statistics and probability theory, linear algebra. Technical Skills: Proficiency in Java (preferred) or another object-oriented language is required, including effective application of design patterns and best coding practices. DevOps experience, with a good command of CI/CD process and tools (eg, Git, GitHub, Gradle, Jenkins, Docker, Helm, Harness). Experience in containerized deployment in cloud environments. Experienced with cloud technology (AWS preferred), infrastructure-as-code (eg Terraform), managing and orchestrating containerized workloads (eg Kubernetes). Education and/or Experience: Master's degree or equivalent in a computational or numerical field such as computer science, information systems, mathematics, physics 7+ years of experience as a software developer with exposure to the cloud or high-performance computing areas
*We are unable to sponsor for this role* *At the minimum year-long contract* A prestigious company is on the search for a Data Quality Test Engineer. The Test Engineer will be responsible for ensuring the successful implementation of new features as well as improvement of existing functionality for data architecture projects. This includes Integration testing, Customer Data Quality testing, Data verification testing, Regression testing etc. Responsibilities Review business requirements, functional specifications, use cases, design, and architecture documents to create effective, value-add test plans. Identify and close gaps, by assisting in QA discovery or analysis of a built or process, in order to facilitate requirements refinements. Identify issues in data, application configurations and screen designs, report and oversee issue resolution. Plan and execute Manual, Functional, and Regression test suite. Independently develop, modify, and execute test plans based on requirements and design documents. Identify, report and track defects and User Experience issues through a system of record. Collaborate with external teams on root cause analysis. Ensuring proper traceability from requirements to test artifacts. Actively participate in requirements and code reviews. Actively participate and drive improvements in Agile Scrum delivery. Be a thought leader in the test design and execution arena (manual and automated). Creation and input of test data, either manually or by using tools, programs, or masked production data to execute test cases. Qualifications: Bachelor's degree in Computer Science or related area. At least 3-5 years of testing experience in data or ETL projects. Strong experience with writing queries in relational databases like Oracle and SQL server. Familiar with SQL server queries including calling stored procedures by passing parameters, functions, views, and indexes to be used during ETL process. Possess extensive and in-depth knowledge of developing SQL scripts using SQL functions, grouping operations, sub queries, analytical function and joins to test ETL projects. ETL testing using SQL server Integration Services (SSIS) and SSRS. Experience in testing database applications to validate source to destination data movement and transformation. Knowledge of automated data testing is preferred. Hands-on experience in Planning and Manual test execution. Experience in Regression Test Strategies, Regression suite maintenance and execution. Hands-on experience with Jira, ALM, TFS/ADO or similar systems of record. Hands-on experience in Agile and Waterfall methodologies.
16/05/2024
Project-based
*We are unable to sponsor for this role* *At the minimum year-long contract* A prestigious company is on the search for a Data Quality Test Engineer. The Test Engineer will be responsible for ensuring the successful implementation of new features as well as improvement of existing functionality for data architecture projects. This includes Integration testing, Customer Data Quality testing, Data verification testing, Regression testing etc. Responsibilities Review business requirements, functional specifications, use cases, design, and architecture documents to create effective, value-add test plans. Identify and close gaps, by assisting in QA discovery or analysis of a built or process, in order to facilitate requirements refinements. Identify issues in data, application configurations and screen designs, report and oversee issue resolution. Plan and execute Manual, Functional, and Regression test suite. Independently develop, modify, and execute test plans based on requirements and design documents. Identify, report and track defects and User Experience issues through a system of record. Collaborate with external teams on root cause analysis. Ensuring proper traceability from requirements to test artifacts. Actively participate in requirements and code reviews. Actively participate and drive improvements in Agile Scrum delivery. Be a thought leader in the test design and execution arena (manual and automated). Creation and input of test data, either manually or by using tools, programs, or masked production data to execute test cases. Qualifications: Bachelor's degree in Computer Science or related area. At least 3-5 years of testing experience in data or ETL projects. Strong experience with writing queries in relational databases like Oracle and SQL server. Familiar with SQL server queries including calling stored procedures by passing parameters, functions, views, and indexes to be used during ETL process. Possess extensive and in-depth knowledge of developing SQL scripts using SQL functions, grouping operations, sub queries, analytical function and joins to test ETL projects. ETL testing using SQL server Integration Services (SSIS) and SSRS. Experience in testing database applications to validate source to destination data movement and transformation. Knowledge of automated data testing is preferred. Hands-on experience in Planning and Manual test execution. Experience in Regression Test Strategies, Regression suite maintenance and execution. Hands-on experience with Jira, ALM, TFS/ADO or similar systems of record. Hands-on experience in Agile and Waterfall methodologies.
Job Description: As a Cassandra Database Administrator, you will be responsible for the design, implementation, and maintenance of our Cassandra database clusters, ensuring their scalability, reliability, and performance to meet the demands of our growing business. You will collaborate closely with software engineers, data architects, and other stakeholders to design efficient data models, optimize query performance, and ensure data consistency and availability. Your primary responsibilities will include: Installing, configuring, and managing Apache Cassandra clusters, including nodes, partitions, replication, and consistency levels, across multiple environments (eg, development, testing, production). Monitoring database health, performance metrics, and resource utilization, and proactively identifying and resolving issues to minimize downtime and ensure optimal performance. Implementing and maintaining data replication, backup, and recovery strategies to safeguard data integrity and availability in the event of failures or disasters. Managing schema design, data modelling, and query optimization to maximize performance, scalability, and efficiency of Cassandra databases. Implementing security measures, access controls, and encryption mechanisms to protect sensitive data and comply with regulatory requirements. Collaborating with cross-functional teams to design and implement data migration, ETL processes, and data integration workflows between Cassandra and other data sources. Performing capacity planning, scaling, and performance tuning to accommodate growing data volumes, user loads, and application requirements. Automating routine tasks, such as cluster provisioning, configuration management, monitoring, and alerting, using Scripting languages and automation tools.
16/05/2024
Project-based
Job Description: As a Cassandra Database Administrator, you will be responsible for the design, implementation, and maintenance of our Cassandra database clusters, ensuring their scalability, reliability, and performance to meet the demands of our growing business. You will collaborate closely with software engineers, data architects, and other stakeholders to design efficient data models, optimize query performance, and ensure data consistency and availability. Your primary responsibilities will include: Installing, configuring, and managing Apache Cassandra clusters, including nodes, partitions, replication, and consistency levels, across multiple environments (eg, development, testing, production). Monitoring database health, performance metrics, and resource utilization, and proactively identifying and resolving issues to minimize downtime and ensure optimal performance. Implementing and maintaining data replication, backup, and recovery strategies to safeguard data integrity and availability in the event of failures or disasters. Managing schema design, data modelling, and query optimization to maximize performance, scalability, and efficiency of Cassandra databases. Implementing security measures, access controls, and encryption mechanisms to protect sensitive data and comply with regulatory requirements. Collaborating with cross-functional teams to design and implement data migration, ETL processes, and data integration workflows between Cassandra and other data sources. Performing capacity planning, scaling, and performance tuning to accommodate growing data volumes, user loads, and application requirements. Automating routine tasks, such as cluster provisioning, configuration management, monitoring, and alerting, using Scripting languages and automation tools.
KDB Developer Hybrid - London Contract Hayward Hawk are currently recruiting for a KDB Developer for their global consultancy client. The successful candidate will be responsible for designing, developing, and maintaining KDB databases, as well as creating and optimizing Real Time data analytics solutions. You will work closely with the technology team to enhance our data management systems, ensure data accuracy and integrity, and help us make data-driven decisions. Key Responsibilities: Design and develop KDB databases and data processing systems to support Real Time data analysis and reporting. Optimize database performance, troubleshoot issues, and implement improvements as needed. Collaborate with data analysts and other teams to understand data requirements and develop efficient data retrieval processes. Write and maintain KDB Q scripts, ensuring code is well-documented and follows best practices. Monitor data quality and accuracy, implementing data validation and cleansing processes. Develop and maintain data pipelines and integration with various data sources and systems. Stay up-to-date with industry best practices and new technologies to ensure our data management systems remain cutting-edge. Collaborate with other developers and technology professionals to drive projects to successful completion. Participate in code reviews and share your knowledge with team members. Provide production support as needed, including resolving any database-related issues and ensuring system availability. Qualifications: 3-5 years of hands-on experience with KDB/Q development. Strong proficiency in Q programming language. Experience with Real Time data analysis and financial systems (preferred). Knowledge of database design principles, performance optimization, and Datamodelling. Familiarity with data integration, ETL processes, and data warehousing. Excellent problem-solving skills and the ability to work effectively in a fast-paced environment. Strong communication and teamwork skills. A proactive and self-motivated approach to work. Familiarity with financial markets and trading data would be advantageous
15/05/2024
Project-based
KDB Developer Hybrid - London Contract Hayward Hawk are currently recruiting for a KDB Developer for their global consultancy client. The successful candidate will be responsible for designing, developing, and maintaining KDB databases, as well as creating and optimizing Real Time data analytics solutions. You will work closely with the technology team to enhance our data management systems, ensure data accuracy and integrity, and help us make data-driven decisions. Key Responsibilities: Design and develop KDB databases and data processing systems to support Real Time data analysis and reporting. Optimize database performance, troubleshoot issues, and implement improvements as needed. Collaborate with data analysts and other teams to understand data requirements and develop efficient data retrieval processes. Write and maintain KDB Q scripts, ensuring code is well-documented and follows best practices. Monitor data quality and accuracy, implementing data validation and cleansing processes. Develop and maintain data pipelines and integration with various data sources and systems. Stay up-to-date with industry best practices and new technologies to ensure our data management systems remain cutting-edge. Collaborate with other developers and technology professionals to drive projects to successful completion. Participate in code reviews and share your knowledge with team members. Provide production support as needed, including resolving any database-related issues and ensuring system availability. Qualifications: 3-5 years of hands-on experience with KDB/Q development. Strong proficiency in Q programming language. Experience with Real Time data analysis and financial systems (preferred). Knowledge of database design principles, performance optimization, and Datamodelling. Familiarity with data integration, ETL processes, and data warehousing. Excellent problem-solving skills and the ability to work effectively in a fast-paced environment. Strong communication and teamwork skills. A proactive and self-motivated approach to work. Familiarity with financial markets and trading data would be advantageous
NO SPONSORSHIP Associate Principal, Software Programming Quantitative Risk Management Area Associate Principal, Software Engineering Automating Risk Models Chicago - On site 3 days a week Salary - $185 - $195K + Bonus Looking for a hard core developer who works within the quantitative risk management and cab develop applications and solutions for the QRM team. You will not build models, you will automate models You will need to come from a financial institute, trading company, exchange, etc. Develop hardcore applications You will need to have CICD pipelines, Infrastructure as a Code, Kubernetes, Terraform, etc. Preferably having Java, Python, C++ Configure and manage resources in the local and AWS cloud environments and deploy QRMs software on these resources. Develop CI/CD pipelines. Contribute to development of QRMs databases and ETLs. Integrate model prototypes, model library and model testing tools using best industry practices and innovations. Create unit and integration tests; build and enhance test automation tools. Participate in code reviews and demo accomplishments. Write technical documentation and user manuals. Provide production support and perform troubleshooting. Strong programming skills. Able to read and/or write code using a programming language (eg, Java, C++, Python, etc.) in a collaborative software development setting: The role requires advanced coding, database and environment manipulation skills. cloud environment. Financial products knowledge is a plus: understanding of markets and financial derivatives in equities, interest rate, and commodity products. Background in Financial mathematics is a plus: derivatives pricing models, stochastic calculus, statistics and probability theory, linear algebra. Technical Skills: Proficiency in Java (preferred) or another object-oriented language is required, including effective application of design patterns and best coding practices. DevOps experience, with a good command of CI/CD process and tools (eg, Git, GitHub, Gradle, Jenkins, Docker, Helm, Harness). Experience in containerized deployment in cloud environments. Experienced with cloud technology (AWS preferred), infrastructure-as-code (eg Terraform), managing and orchestrating containerized workloads (eg Kubernetes). Education and/or Experience: Masters degree or equivalent in a computational or numerical field such as computer science, information systems, mathematics, physics 7+ years of experience as a software developer with exposure to the cloud or high-performance computing areas
14/05/2024
Full time
NO SPONSORSHIP Associate Principal, Software Programming Quantitative Risk Management Area Associate Principal, Software Engineering Automating Risk Models Chicago - On site 3 days a week Salary - $185 - $195K + Bonus Looking for a hard core developer who works within the quantitative risk management and cab develop applications and solutions for the QRM team. You will not build models, you will automate models You will need to come from a financial institute, trading company, exchange, etc. Develop hardcore applications You will need to have CICD pipelines, Infrastructure as a Code, Kubernetes, Terraform, etc. Preferably having Java, Python, C++ Configure and manage resources in the local and AWS cloud environments and deploy QRMs software on these resources. Develop CI/CD pipelines. Contribute to development of QRMs databases and ETLs. Integrate model prototypes, model library and model testing tools using best industry practices and innovations. Create unit and integration tests; build and enhance test automation tools. Participate in code reviews and demo accomplishments. Write technical documentation and user manuals. Provide production support and perform troubleshooting. Strong programming skills. Able to read and/or write code using a programming language (eg, Java, C++, Python, etc.) in a collaborative software development setting: The role requires advanced coding, database and environment manipulation skills. cloud environment. Financial products knowledge is a plus: understanding of markets and financial derivatives in equities, interest rate, and commodity products. Background in Financial mathematics is a plus: derivatives pricing models, stochastic calculus, statistics and probability theory, linear algebra. Technical Skills: Proficiency in Java (preferred) or another object-oriented language is required, including effective application of design patterns and best coding practices. DevOps experience, with a good command of CI/CD process and tools (eg, Git, GitHub, Gradle, Jenkins, Docker, Helm, Harness). Experience in containerized deployment in cloud environments. Experienced with cloud technology (AWS preferred), infrastructure-as-code (eg Terraform), managing and orchestrating containerized workloads (eg Kubernetes). Education and/or Experience: Masters degree or equivalent in a computational or numerical field such as computer science, information systems, mathematics, physics 7+ years of experience as a software developer with exposure to the cloud or high-performance computing areas
Lead Data Engineer, Informatica Liverpool/Hybrid Circa £110k + bonus + benefits Databricks, Informatica, Azure Lead/Senior Data Engineer is required by vibrant Insurance organisation with offices in Liverpool. This is a key role within their data team that will see you driving forward new exciting data initiatives. You will help the organisation to implement what is needed to deliver strategic data platforms. This will include, but not limited to the end-to-end support for acquisition, organisation, integration, transformation and secure storage of mission-critical data. Key Responsibilities: Work closely in partnership with large system integrators to support the delivery of the programs that align to data strategy. Work with one or more delivery teams on a given program to ensure requirements are obtained, work has been properly identified, and lead or support engineers in delivering the data products required as part of the programs. Partner with both the business and technical team to ensure that solutions are aligned to the business needs and follow enterprise standards and guidelines. Partner with data management and governance leads to ensure that projects are delivered at or under budget and follow all compliance, regulatory, and data quality requirements. Mentor the Data Engineers in the team. Key Requirements: Significant experience and in-depth knowledge of data delivery; Datamodelling concepts, ETL procedures, and all steps of data production process Extensive hands-on experience in Informatica IICS, ADF, Warehousing technologies and associated patterns, Cloud platforms - Azure preferred Analyses data requirements and provides data analysis techniques, and applies Datamodelling (including data vault) and data quality techniques to establish, modify or maintain data structures and their associated components in complex environments Experience with Agile delivery frameworks/methodologies (eg Scrum, SAFe) and tools (ie Jira, Azure DevOps) Demonstrate a proven track record of having implemented solutions using one or more agile methodologies in conjunction with a large system integration partner. Experience with mass ingestion capabilities and cloud process flows, data quality and master data management Understanding data related security challenges and tooling with specific technologies (eg Databricks) Experience in migrating data and associated systems from on-prem to cloud environments Consulting and collaborative working style creating a culture of accountability and sharing Knowledge of advanced analytics (models, tools, etc.) and their operational support is a plus *All salaries are subject to experience* For a full consultation on this role please contact Arc IT or email your CV to Arc IT Recruitment. Arc is an equal opportunities recruiter and we welcome applications from all suitably skilled or qualified applicants, regardless of their race, sex, disability, religion/beliefs, sexual orientation or age.
14/05/2024
Full time
Lead Data Engineer, Informatica Liverpool/Hybrid Circa £110k + bonus + benefits Databricks, Informatica, Azure Lead/Senior Data Engineer is required by vibrant Insurance organisation with offices in Liverpool. This is a key role within their data team that will see you driving forward new exciting data initiatives. You will help the organisation to implement what is needed to deliver strategic data platforms. This will include, but not limited to the end-to-end support for acquisition, organisation, integration, transformation and secure storage of mission-critical data. Key Responsibilities: Work closely in partnership with large system integrators to support the delivery of the programs that align to data strategy. Work with one or more delivery teams on a given program to ensure requirements are obtained, work has been properly identified, and lead or support engineers in delivering the data products required as part of the programs. Partner with both the business and technical team to ensure that solutions are aligned to the business needs and follow enterprise standards and guidelines. Partner with data management and governance leads to ensure that projects are delivered at or under budget and follow all compliance, regulatory, and data quality requirements. Mentor the Data Engineers in the team. Key Requirements: Significant experience and in-depth knowledge of data delivery; Datamodelling concepts, ETL procedures, and all steps of data production process Extensive hands-on experience in Informatica IICS, ADF, Warehousing technologies and associated patterns, Cloud platforms - Azure preferred Analyses data requirements and provides data analysis techniques, and applies Datamodelling (including data vault) and data quality techniques to establish, modify or maintain data structures and their associated components in complex environments Experience with Agile delivery frameworks/methodologies (eg Scrum, SAFe) and tools (ie Jira, Azure DevOps) Demonstrate a proven track record of having implemented solutions using one or more agile methodologies in conjunction with a large system integration partner. Experience with mass ingestion capabilities and cloud process flows, data quality and master data management Understanding data related security challenges and tooling with specific technologies (eg Databricks) Experience in migrating data and associated systems from on-prem to cloud environments Consulting and collaborative working style creating a culture of accountability and sharing Knowledge of advanced analytics (models, tools, etc.) and their operational support is a plus *All salaries are subject to experience* For a full consultation on this role please contact Arc IT or email your CV to Arc IT Recruitment. Arc is an equal opportunities recruiter and we welcome applications from all suitably skilled or qualified applicants, regardless of their race, sex, disability, religion/beliefs, sexual orientation or age.
*Hybrid, 3 days onsite, 2 days remote* *We are unable to sponsor as this is a permanent Full time role* A prestigious company is looking for a Director, Software Engineering - QRM. This director will manage 6 people and will help develop software applications and solutions for the quantitative management platform. This director will need hands-on experience with Java, DevOps, CICD, AWS, Containers, terraform, Etc. Responsibilities: Develop and maintain software and environments used to implement and test systems for pricing, margin risk and stress testing of financial products and derivatives. Configure and manage resources in the local and AWS cloud environments and deploy QRM's software on these resources. Develop CI/CD pipelines. Configure, execute, and monitor execution pipelines for model testing, backtesting and monitoring. Contribute to development of QRM's databases and ETLs. Integrate model prototypes, model library and model testing tools using best industry practices and innovations. Create unit and integration tests; build and enhance test automation tools. Participate in code reviews and demo accomplishments. Write technical documentation and user manuals. Provide production support and perform troubleshooting. Provide hands-on technical leadership and active coordination of tasks and priorities. Provide guidance and support for the team and reporting for the management. Qualifications: Master's degree or equivalent in a computational or numerical field such as computer science, information systems, mathematics, physics 10+ years of experience as a software developer with exposure to the cloud or high-performance computing areas Strong programming skills. Able to read and/or write code using a programming language (eg, Java, C++, Python, etc.) in a collaborative software development setting: The role requires advanced coding, database and environment manipulation skills. Proficiency in Java (preferred) or another object-oriented language is required, including effective application of design patterns and best coding practices. DevOps experience, with a good command of CI/CD process and tools (eg, Git, GitHub, Gradle, Jenkins, Docker, Helm, Harness). Experience in containerized deployment in cloud environments. Experienced with cloud technology (AWS preferred), infrastructure-as-code (eg Terraform), managing and orchestrating containerized workloads (eg Kubernetes). Experience with logging, profiling, monitoring, telemetry (eg Splunk, OpenTelemetry). Good command of database technology and query languages (SQL) and non-relational DB and other Big Data technology, including efficient storage and serialization protocols (eg Parquet, Avro, Protocol Buffers). Experience with automated quality assurance frameworks (eg, Junit, TestNG, PyTest, etc.). Experience with high performance and distributed computing. Experience with productivity tools such as Jira, Confluence, MS Office.
10/05/2024
Full time
*Hybrid, 3 days onsite, 2 days remote* *We are unable to sponsor as this is a permanent Full time role* A prestigious company is looking for a Director, Software Engineering - QRM. This director will manage 6 people and will help develop software applications and solutions for the quantitative management platform. This director will need hands-on experience with Java, DevOps, CICD, AWS, Containers, terraform, Etc. Responsibilities: Develop and maintain software and environments used to implement and test systems for pricing, margin risk and stress testing of financial products and derivatives. Configure and manage resources in the local and AWS cloud environments and deploy QRM's software on these resources. Develop CI/CD pipelines. Configure, execute, and monitor execution pipelines for model testing, backtesting and monitoring. Contribute to development of QRM's databases and ETLs. Integrate model prototypes, model library and model testing tools using best industry practices and innovations. Create unit and integration tests; build and enhance test automation tools. Participate in code reviews and demo accomplishments. Write technical documentation and user manuals. Provide production support and perform troubleshooting. Provide hands-on technical leadership and active coordination of tasks and priorities. Provide guidance and support for the team and reporting for the management. Qualifications: Master's degree or equivalent in a computational or numerical field such as computer science, information systems, mathematics, physics 10+ years of experience as a software developer with exposure to the cloud or high-performance computing areas Strong programming skills. Able to read and/or write code using a programming language (eg, Java, C++, Python, etc.) in a collaborative software development setting: The role requires advanced coding, database and environment manipulation skills. Proficiency in Java (preferred) or another object-oriented language is required, including effective application of design patterns and best coding practices. DevOps experience, with a good command of CI/CD process and tools (eg, Git, GitHub, Gradle, Jenkins, Docker, Helm, Harness). Experience in containerized deployment in cloud environments. Experienced with cloud technology (AWS preferred), infrastructure-as-code (eg Terraform), managing and orchestrating containerized workloads (eg Kubernetes). Experience with logging, profiling, monitoring, telemetry (eg Splunk, OpenTelemetry). Good command of database technology and query languages (SQL) and non-relational DB and other Big Data technology, including efficient storage and serialization protocols (eg Parquet, Avro, Protocol Buffers). Experience with automated quality assurance frameworks (eg, Junit, TestNG, PyTest, etc.). Experience with high performance and distributed computing. Experience with productivity tools such as Jira, Confluence, MS Office.