--- license: mit --- # Checkpoint for Multistain Pretraining for Slide Representation Learning in Pathology (ECCV'24) Welcome to the official HuggingFace repository of the ECCV 2024 paper, ["Multistain Pretraining for Slide Representation Learning in Pathology"](https://huggingface.co/papers/2408.02859). This project was developed at the [Mahmood Lab](https://faisal.ai/) at Harvard Medical School and Brigham and Women's Hospital. ## Model loging ``` from huggingface_hub import notebook_login notebook_login() ``` You can refer [HuggingFace](https://huggingface.co/docs/huggingface_hub/en/quick-start#login-command) documentation for more details. ## Preprocessing: tissue segmentation, patching, and patch feature extraction We are extracting [CONCH](https://github.com/mahmoodlab/CONCH) features at 10x magnification on 256x256-pixel patches. Please refer to Madeleine public implementation to extract patch embeddings from a WSI. ## Extracting MADELEINE slide encoding You can obtain and run MADELEINE slide encoding (trained on Acrobat breast samples at 10x magnification) using: ``` from core.models.factory import create_model_from_pretrained model, precision = create_model_from_pretrained('./models/') feats = load_h5('your_path_to_conch_patch_embeddings/XXX.h5') with torch.no_grad(): with torch.amp.autocast(device_type="cuda", dtype=precision): wsi_embed = model.encode_he(feats=feats, device='cuda') ```