- Experience: 5 + Years
- Location: Lahore – Onsite
- Working Hours: 5:00 PM to 2:00 AM (Monday to Friday)
Primary Responsibilities:
- Design and implement end-to-end machine learning solutions to process and analyze large-scale datasets efficiently·
- Build robust data pipelines for high-volume data ingestion, transformation, and storage·
- Create advanced reporting systems and dashboards powered by AI to deliver actionable insights·
- Develop and deploy custom AI/ML models tailored to business requirements, including LLMs and generative AI solutions·
- Leverage frameworks like LangChain, LlamaIndex, and Hugging Face to implement and fine-tune large language models·
- Build systems to automate data-driven decision-making using predictive analytics and recommendation engines·
- Monitor, troubleshoot, and optimize AI/ML models in production environments·
- Collaborate with cross-functional teams to identify opportunities to enhance reporting and analytics capabilities with AI·
- Ensure scalability, reliability, and security in handling vast amounts of data·
- Maintain comprehensive documentation for data workflows, AI systems, and reporting tools
Technical Requirements:
Bachelor’s, Master’s, or PhD in a STEM field (Science, Technology, Engineering, or Mathematics)·
- 5+ years of experience in machine learning and large-scale data processing with expertise in:
- Python, SQL, and data processing frameworks (Pandas, NumPy, PySpark, Dask)
- Big data tools (Hadoop, Apache Spark, or similar)
- Machine learning frameworks (TensorFlow, PyTorch, or similar)
- Building and deploying large language models (LLMs) and generative AI systems
- REST APIs and microservices architecture for scalable AI solutions
- Version control (Git) and CI/CD pipelines for MLOps
- Data visualization tools (Tableau, Power BI, or similar)
- Strong experience with:
- Real-time and batch processing of large datasets
- Model optimization and monitoring in production environments
- NLP and generative AI techniques for text analysis and automation
- Feature engineering, selection, and time series forecasting
- Database management (SQL and NoSQL)
Required Skills:
- Proven ability to process, analyze, and report on large-scale datasets·
- Strong mathematical and statistical foundation·
- Expertise in building data-driven reporting systems and tools·
- Hands-on experience with LLM optimization and generative AI pipelines·
- Knowledge of cloud platforms (AWS, Google Cloud, or Azure) for scalable data processing·
- Strong communication skills to collaborate with data teams and business stakeholders·
- Understanding of data security and compliance best practices
Preferred Qualifications:
- Experience with distributed computing and big data architectures·
- Familiarity with LangChain, LlamaIndex, Hugging Face, and similar LLM-focused tools·
- Knowledge of A/B testing and experimental design for data-driven insights·
- Proficiency in containerization and orchestration (Docker, Kubernetes)·
- Published research or contributions to AI/ML open-source projects·
- Experience with advanced model optimization and hardware acceleration
Skills:
NumPy, SQL, Hadoop, Apache, Restful APIs, Microservices, Machine Learning, Data Processing, Python Framework Command, Panda, Data Visualization Skills, PySpark, Dask, Big Data, Tensoflow, PyTorch, LLMs, AI Solutions, Git, Pipeline Management,