Senior Software Engineer- Serverless Infrastructure for Data Science | New York, NY | Bloomberg Careers

Senior Software Engineer- Serverless Infrastructure for Data Science

Careers at Bloomberg

Back to Search

New York, NY

Posted Nov 10, 2022 - Requisition No. 108998

Bloomberg runs on data. It's our business and our product. From the biggest banks to elite hedge funds, financial institutions need timely, accurate data to capture opportunities and evaluate risk in fast-moving markets. With petabytes of data available, a platform to transform and analyze the data is critical to our success. 

Bloomberg’s Data Science Platform was established to support development efforts around data-driven science, machine learning, and business analytics on Bloomberg's many datasets.

The platform was developed to provide a standard set of tooling for the Model Development Life Cycle (MDLC) which includes tools for the early stages of development and data exploration, to experimentation and large scale training, all the way to live inference. Through access to scalable compute and specialized hardware, Data Science Platform users have access to ML training jobs and Inference Services, Analytics and ETL using Spark, and data exploration with Jupyter. The platform is built on Kubernetes, leveraging containerization, container orchestration and a cloud architecture using 100% open source foundations.

What We do:

Model Prediction, or Inference, is the last critical step in the MDLC,  when business value of model-driven applications can be realized. Our Inference solution is powered by the open source project KServe, a highly scalable and standards based model inference platform for trusted AI. 

Delivering performance to latency-sensitive, throughput-heavy, model-driven applications, means making the right choices from hardware to Ingress.  As a member of the Data Science Platform’s Infrastructure team with a focus on the Serverless Components, you’ll have the opportunity to work on open source serverless technologies underlying KServe, such as Knative and Istio, as well as looking at the latest hardware available on the market to service hundreds to thousands of models in a scalable way.

As the founding members of KServe, we regularly upstream features we develop, present at conferences and collaborate with our peers in the industry, and are in tune with the surrounding Kubernetes community. Open source is at the heart of our team. It's not just something we do in our free time, it is how we work.

We’ll trust you to:

  • Innovate and design solutions that keep in mind strict production SLA: low latency/high throughput, multi-tenancy, high availability, reliability across clusters/data centers, etc.
  • Interact with ML experts to understand workflows, pinpoint and resolve inefficiencies, and inform the next set of features for the platform.
  • Collaborating with open-source communities and internal platform teams to build cohesive model deployment experience.
  • Automate operations and improve observability of the platform by integrating with systems for metrics and distributed tracing.
  • Troubleshoot and optimize ML model inference performance.
  • Build tools enabling other engineers a way to debug and understand performance of complicated systems.

What we are looking for: 

  • Have a passion for providing reliable and scalable ML infrastructure
  • Experience designing and implementing low-latency, high-scalability systems
  • Experience working in a multi-tenancy and multi-cluster environment
  • Experience with ML infrastructure open source project such as Kubeflow, KServe, MLFlow, Feast
  • Experience with distributed systems eg. Kubernetes, Kafka, Zookeeper/Etcd, Spark
  • Experience with debugging performance issues with distributed tracing and benchmark tools
  • Proficiency in two or more languages (Go, Python, C++, or JavaScript) and willingness to learn new technologies as needed
  • At least 2 years of experience as a software engineer

We’d love to see: 

  • Experience with serverless framework or infrastructure, such as Knative, AWS Lambda, Google CloudRun.
  • Experience working with Service Mesh, authentication & authorization systems like Spire/Spiffe.
  • Experience working with GPU compute software and hardware
  • Ability to identify and perform OS and hardware-level optimizations
  • Open source involvement such as a well-curated blog, accepted contribution, or community presence
  • Experience with cloud providers such as AWS, GCP or Azure
  • Experience with configuration management systems (Chef, Puppet, Ansible, or Salt)
  • Experience with continuous integration tools and technologies (Jenkins, Git, Chat-ops)
  • Passion for education e.g providing workshops for tenants

If this sounds like you, apply! You can also learn more about our work using the links below:

We are an equal opportunity employer and value diversity at our company. We do not discriminate on the basis of race, religion, color, national origin, gender, sexual orientation, age, marital status, veteran status, or disability status.

Salary Range: 160,000 - 240,000 USD Annually + Benefits + Bonus

The referenced salary range is based on the Company's good faith belief at the time of posting. Actual compensation may vary based on factors such as geographic location, work experience, market conditions, education/training and skill level.

We offer one of the most comprehensive and generous benefits plans available and offer a range of total rewards that may include merit increases, incentive compensation [Exempt roles only], paid holidays, paid time off, medical, dental, vision, short and long term disability benefits, 401(k) +match, life insurance, and various wellness programs, among others. The Company does not provide benefits directly to contingent workers/contractors and interns.

Similar jobs

    The Bloomberg Talent Network

    Stay connected with us and be among the first to learn about new job opportunities. We’ll use the information you provide to help us get in touch with you to align your expertise with our opportunities and better direct our conversations.