How can I help?

I am a Machine Learning Engineer that specialises in generative AI & LLMs. Having spent a number of years working as a full time data engineer and ML engineer, I am now self-employed and work on a contract basis. I have a masters degree in computer science with a specialisation in data mining.

My contracts usually involve architecture, development, teaching or auditing. I enjoy working independently and as part of a team.

5

years of experience

15+

clients

100%

satisfaction rate

Please get in contact on linkedin if you'd like to explore opportunities to work together.

Contact Portfolio

What can I do for you?

Data management

Manage and clean your company's data, integrate external sources and expose it to other applications, such as business intelligence, with ETL pipelines.

Datascience

AI Pipelines

Build AI pipelines augmented with your company's data (RAG),

Micro services

Architecture and deploy resilient micro services for your organization through Docker/Cloud/APIs.

CI/CD

Automate, speed-up and optimize the integration, delivery and deployment of your applications.

Skillset

I am a back-end systems specialist and data expert; however, I also have experience as a full stack engineer and can pick up new technologies quickly.

My mission is to make sure your organization's data is well prepared, clean and reliable. I can help you create large scale data pipelines that may involve bringing many different systems together.

As a data engineer, I can help you choose the right tools for the job, and combine them to create solutions to enable your company’s business processes with data pipelines.

Check out how I helped MyTraffic in collecting and leveraging pedestrian footfall data to predict best retail locations...
Check out the big data pipelines I developed for the BI team at Gumgum....

Languages

Frameworks
Apache Spark logo

My mission is to help you understand if and where machine learning can be useful for your company, and implement the AI models and systems for you.

As a data scientist, I can help you create data pipelines for machine learning (extraction, cleaning, augmentation) and deploy state of the art models suited to your problematic.

Check out my work on fake faces recognition on a 500GB video dataset…
Check out my work on building a bot detection engine… Check out my open-source image classification app…

Languages

Frameworks
Keras Logo

I can work with your team to shorten and optimize your application lifecycle process, and facilitate fast and clean deployments via standardization and automation. I am also able to architect your systems using a cloud provider of your choice (AWS, GCP, Azure, IBMCloud)

If you need to deploy CI/CD pipelines across your organization or expose applications as micro-services, on-premises or on the cloud, I can do it for you.

Languages
Frameworks
Docker Logo

As a ML engineer, I bridge the gap between data scientists, data engineers and devops engineers.
I help you on the following tasks :

  • Design, architect and deploy end-to-end AI systems.
  • Develop the MLOps culture and principles in your company.
  • Improve model deployment, training and monitoring processes
  • Scale existing models, reduce time to market for AIs
  • Evaluate and deploy MLOps tools (MLFlow, DVC, W&B, Kubeflow...)

  • ml_workflow_oreilly
    Source : Introducing MLOps - O'reilly ®
    Frameworks
    AWS Sagemaker Logo Metaflow Logo Kubeflow Logo

    Recommendations

    Latest work

    Here are some examples of projects I worked on, in various contexts (start-ups, bigger companies, independently, as an employee).

    Contact me Resume


    MLOps architecture

    At ManoMano, I designed the architecture of an MLOps platform on AWS, then implememented and industrialized for all the data teams in the company. It is currently used daily by the data science team for building, deploying and monitoring ML models.

    Feb 2022
    Data training

    I teach data courses, trainings in schools and companies (Model deployment, data science, data engineering, CI/CD, best practices in data)

    Apr 2022

    At MyTraffic, I implemented data science models at high scale, and deployed them on the cloud for pedestrian traffic insights.

    Jan 2020
    Big data pipeline at GumGum

    Built and maintained batch and real-time data pipelines to provide business insights on a continuous basis. These pipelines were downstream from a programmatic ad-serving platform generating 50 TB of data a day.

    Feb 2019
    Cruise control architecture

    Deployed a suite of monitoring and management tools across multiple Kafka clusters. These tools reduced the engineering time of certain cluster operations (scaling, rebalancing) by 90% as well as enhanced the overall visibility into the health of these clusters.

    Apr 2019
    Thumbnail [100%x225]

    Developed a high-scale programmatic inventory forecasting web tool. This tool uses an ARIMA model and terabytes of historical data to predict ad inventory. In close cooperation with operational teams, we provided the end-users the ability to determine achievable ad campaigns goals with confidence using the forecast results.

    Thumbnail [100%x225]

    A simple computer vision system able to classify images into 1000 object categories, such as keyboard, pencil, and many animals. Visit the UI or try it out below.

    Feb 2020

    Deeplabv3+

    DeepLab is a state-of-art deep learning model for semantic image segmentation, where the goal is to assign semantic labels (e.g., person, car, dog, bicycle and so on) to every pixel in the input image.

    DeepLabv3+, extends DeepLabv3 by adding a simple yet effective decoder module to refine the segmentation results especially along object boundaries.

    According to Papers With Code, it has the best accuracy and is the current state of the art on the Pascal VOC dataset.

    The code is available here.

    Try it out !

    Or send one of the sample images...

    Click to send
    Click to send
    Click to send