Oracle
Senior Data Engineer – AI Accelerator
Job Description
The Infrastructure Industries Global Industry Unit’s (IGIU) mission is to build and deliver technology and solutions that improve the lives of global citizens. For Energy & Water, that’s striving to ensure that every global citizen has access to clean and affordable energy and water. For Construction & Engineering, it’s navigating enormous change to become more efficient, smarter and sustainable.
You’ll be joining a diverse, cross-functional team that’s developing an Oracle Cloud Infrastructure (OCI) based AI powered platform for Infrastructure industries. We operate with a startup mindset providing the best of two worlds – the autonomy to iterate on new ideas, working on cutting-edge technology, while sitting within a stable, larger organization at Oracle. We’re working on big goals, and we need talented folks with equally big ambitions. Join us!
We’re seeking a well-rounded Senior Data Engineer to design and build the data platform that will enable IGIU’s product portfolio to quickly uptake OCI AI capabilities, with a focus on Generative AI. You will collaborate closely with cross-functional teams including application product managers, software engineers, and data scientists. Our new platform will be built directly on OCI based on cloud native principles. We build to scale globally, leveraging state-of-the-art tooling, with zero downtime.
You will be responsible for preparing data for ingest into AI models, building scalable AI pipelines and developing ML/LLMOps capabilities. You will become an expert on data ingestion, curation, management and storage for AI model development and testing.
Career Level – IC3
- Contribute to the development a catalog of data across the Infrastructure Industries to enable common data access
- Prepare and curate data for ingest into AI models
- Develop the software primitives for the application teams we serve to develop scalable AI pipelines
- Collaborate with engineers to integrate machine learning models into production systems, ensuring scalability, reliability, and performance
- Develop ML/LLMOps capabilities to ensure the accuracy and reliability of our AI pipelines.
- Utilize vector databases and advanced indexing techniques to efficiently store and retrieve relevant information for conversational contexts
About You
- Excellent problem-solving abilities and a pragmatic approach to building scalable and robust machine learning systems.
- Strong communication skills with the ability to collaborate effectively with cross-functional teams and articulate complex technical concepts to non-technical stakeholders
- You are comfortable with ambiguity. You have a strong sense of ownership, can define your own workplan, and can drive projects to completion.
- You are excited to learn new technologies and stay on the cutting edge of what’s possible
- You’ve taken a product or platform from 0 to 1, you know what it takes to launch something novel
- Bachelor’s degree in Computer Science, Mathematics, related technical field, or equivalent practical experience.
- Experience with Python and/or Java
- Knowledge of the SQL Dialect
- Experience with modern big data processing technologies such as Apache Spark
- Experience developing, maintaining and supporting production grade distributed data processing applications on Cloud Native Stacks like OCI, AWS or Azure
- Strong grasp of system design fundamentals and distributed systems architectural best practices
- Experience in machine learning libraries such as TensorFlow, PyTorch, or JAX.
- Experience in developing RAG models and working with vector databases
#LI-REMOTE
EWJP2