Machine Learning Intern
Onsite
[
New Delhi
]
About Us
At Binaire, we build multi-modal inference systems and frontier text and image models that power next-generation applications. As a fast-growing tech company, we combine cutting-edge research with product-driven engineering to deliver scalable, AI-powered solutions. Join us as a Machine Learning Intern and contribute to real-world ML systems — from model training to deployment.
Role & Responsibilities
As a Machine Learning Intern, you will:
Build & Train Models
- Design, implement, and train machine learning models (e.g., supervised, unsupervised, or deep learning).
- Preprocess data, engineer features, and conduct experiments to improve model accuracy.
Benchmark & Test
- Define, run, and maintain benchmarking pipelines to compare model performance (e.g., latency, accuracy, memory).
- Evaluate models on different datasets and quality metrics.
- Create new benchmarks for key use-cases, to measure model behavior under diverse conditions.
Deploy & Inference
- Deploy models to production or test environments for inference.
- Work with multiple inference engines (e.g., ONNX Runtime, TensorRT, TFLite, or similar) to test model compatibility and performance.
- Optimize models for inference (quantization, pruning, batching).
Application Integration
- Build prototype applications (web or backend) that integrate machine learning models.
- Use web technologies (JavaScript, Node.js, HTML, CSS) to create front-ends or APIs for serving predictions.
Reporting & Documentation
- Create detailed reports on model training results, benchmarking outcomes, and deployment performance.
- Document your experiments, code, and architecture decisions for internal teams and stakeholders.
Research & Innovation
- Explore new techniques, architectures, or inference frameworks.
- Propose and implement improvements to the model lifecycle, such as faster inference, better accuracy, or resource efficiency.
Qualifications & Skills
Must-haves:
- Currently pursuing or recently completed a Bachelor’s or Master’s degree in Computer Science, Data Science, Electrical Engineering, or a related field.
- Strong programming skills in C++ & Python, with experience in ML frameworks (e.g., TensorFlow, PyTorch, scikit-learn).
- Familiarity with web technologies: JavaScript, Node.js, HTML, CSS.
- Understanding of model deployment and inference: knowledge (or willingness to learn) of at least one inference engine/framework (e.g., ONNX, TensorRT, TF-Lite).
- Good problem-solving skills, and ability to run experiments and analyze results.
- Strong written communication skills — ability to write clear reports and document code.
Nice-to-haves:
- Experience with GPU programming or hardware-aware optimizations.
- Exposure to MLOps or model serving tools (e.g., Docker, Kubernetes, MLflow).
- Knowledge of benchmarking tools and metrics for ML (latency, throughput, memory).
- Prior experience building simple web apps or APIs.
- Familiarity with version control (Git) and collaborative workflows.
What You’ll Get
- Hands-on experience in the entire ML lifecycle: from data to deployment
- Mentorship by senior ML engineers and researchers
- Opportunity to contribute to real products and production-grade systems
- Exposure to inference engine optimization
- A chance to present your findings and build benchmarks that guide future work
- Potential for full-time conversion depending on performance