Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

Guide Labs

company
https://www.guidelabs.ai
guidelabsai
https://github.com/guidelabs
Activity Feed

AI & ML interests

We build interpretable models and AI systems that can reliably explain their reasoning, and are easy to audit, steer, and understand.

Andreas Madsen's profile picture Muawiz Chaudhary's profile picture Julius Adebayo's profile picture Isaac Plant's profile picture Giang Nguyen's profile picture

juliusadebayo 
updated a Space about 1 year ago
Running

README

🏃

andreasmadsen 
authored a paper over 1 year ago

Interpretability Needs a New Paradigm

Paper • 2405.05386 • Published May 8, 2024 • 5
andreasmadsen 
authored 6 papers almost 2 years ago

Evaluating the Faithfulness of Importance Measures in NLP by Recursively Masking Allegedly Important Tokens and Retraining

Paper • 2110.08412 • Published Oct 15, 2021 • 1

Faithfulness Measurable Masked Language Models

Paper • 2310.07819 • Published Oct 11, 2023

Measuring Arithmetic Extrapolation Performance

Paper • 1910.01888 • Published Oct 4, 2019

Neural Arithmetic Units

Paper • 2001.05016 • Published Jan 14, 2020

Post-hoc Interpretability for Neural NLP: A Survey

Paper • 2108.04840 • Published Aug 10, 2021

Can Large Language Models Explain Themselves?

Paper • 2401.07927 • Published Jan 15, 2024 • 4
Company
TOS Privacy About Careers
Website
Models Datasets Spaces Pricing Docs