- πββοΈ My name is JoΓ£o Lages
- π·β Deep Learning Engineer @ Revolut
- π± Iβm interested in everything about machine learning, with focus on deep learning applied to text, images, tabular data, video, speech, time-series, anything!
- π« How to reach me: [email protected]
- βοΈ Blog posts:
- Model Merging: MoE, Frankenmerging, SLERP, and Task Vector Algorithms π§ - Deep dive on how LLM merging methods work (co-authored with Deci AI)
- OpenAI JSON Mode vs Functions - Practical differences between these two ways of using OpenAI API
- Direct Preference Optimization (DPO) - A simplified explanation of the DPO algorithm applied to large language models, like Zephyr
- Reinforcement Learning from Human Feedback (RLHF) πββοΈ - A simplified explanation of the RLHF algorithm applied to large language models, like ChatGPT
- Transformers KV Caching Explained πΎ - A short writing on how Key and Value states are cached in transformers for faster inference
- Transformers Positional Encodings Explained π - Positional encoding and how it limits the input size of language models
- Mahalanobis for outlier detection - A simple demo on how to use mahalanobis distance for outlier detection
- β Main open-source contributions:
- Diffusers-Interpret π€π§¨π΅οΈββοΈ - Own package, a model explainability tool built on top of π€ Diffusers
- Ecco - Major contributions to this package that is used to explain, analyze, and visualize NLP language models
- AI Reading Group - Co-author of an open AI reading group from 2019-2023
- RATransformers π - Own package, used to make transformer models relation-aware
- π€ datasets - implemented the mahalanobis distance metric
π
Probably fixing bugs
I live my life as a gradient descent algorithm: one step at a time to find local minimas that maximize my goals.
-
Revolut
- in/thejoaolages
Pinned Loading
-
diffusers-interpret
diffusers-interpret PublicDiffusers-Interpret π€π§¨π΅οΈββοΈ: Model explainability for π€ Diffusers. Get explanations for your generated images.
-
RATransformers
RATransformers PublicRATransformers π- Make your transformer (like BERT, RoBERTa, GPT-2 and T5) Relation Aware!
-
jalammar/ecco
jalammar/ecco PublicExplain, analyze, and visualize NLP language models. Ecco creates interactive visualizations directly in Jupyter notebooks explaining the behavior of Transformer-based language models (like GPT2, Bβ¦
-
Reinforcement Learning from Human Fe...
Reinforcement Learning from Human Feedback (RLHF) - a simplified explanation 1**Maybe you've heard about this technique but you haven't completely understood it, especially the PPO part. This explanation might help.**
23We will focus on text-to-text language models π, such as GPT-3, BLOOM, and T5. Models like BERT, which are encoder-only, are not addressed.
45Reinforcement Learning from Human Feedback (RLHF) has been successfully applied in ChatGPT, hence its major increase in popularity. π
-
Mahalanobis distance: Measure the di...
Mahalanobis distance: Measure the distance of your datapoint to a list of datapoints! 1<!DOCTYPE html>
2<html>
3<head><meta charset="utf-8" />
4<meta name="viewport" content="width=device-width, initial-scale=1.0">
5
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.