Media Summary: EuroPython 2025 — South Hall 2B on 2025-07-17] * How can we reverse engineer what a neural network is doing? In this IASEAI '25 session, An A discussion on the philosophy of deep learning,

Hacking Llms An Introduction To Mechanistic Interpretability Jenny Vega - Detailed Analysis & Overview

EuroPython 2025 — South Hall 2B on 2025-07-17] * How can we reverse engineer what a neural network is doing? In this IASEAI '25 session, An A discussion on the philosophy of deep learning, In this deeply informative video, Jeremy Howard, co-founder of fast.ai and creator of the ULMFiT approach on which all modern ... ACL SIG-FinTech x TFAI Webinar Series ( Understanding and improving Take your personal data back with Incogni! Use code WELCHLABS at the link below and get 60% off an annual plan: ...

In this AI Research Roundup episode, Alex discusses the paper: 'Exploration AI models are trained and not directly programmed, so we don't understand how they do most of the things they do. Our new ... A 4 Step Prompt Injection Methodology - Greet and Repeat While tons of courses dive into pentesting GenAI-powered ... Explore the podcast → AI tools can turn a team of three developers into a fully functioning company. Presentation video for our paper “Jailbreaking Embodied Big thank you to Cisco for sponsoring this video and sponsoring my trip to Cisco Live Amsterdam. // FREE Ethical

Photo Gallery

Hacking LLMs: An Introduction to Mechanistic Interpretability — Jenny Vega
An Introduction to Mechanistic Interpretability – Neel Nanda | IASEAI 2025
Mechanistic Interpretability and How LLMs Understand
A Hackers' Guide to Language Models
Understanding and improving LLMs through mechanistic interpretability
The Dark Matter of AI [Mechanistic Interpretability]
Mechanistic Interpretability for NLP: One-stop Guide for Everything you Need to Know
Understanding and improving LLMs through mechanistic interpretability
Exploration Hacking: LLMs Resisting RL Training
Tracing the thoughts of a large language model
Prompt Injection Methodology for GenAI Application Pentesting - Greet & Repeat Method
LLMjacking: How hackers steal your AI API keys and stick you with the bill
Sponsored
Sponsored
View Detailed Profile
Hacking LLMs: An Introduction to Mechanistic Interpretability — Jenny Vega

Hacking LLMs: An Introduction to Mechanistic Interpretability — Jenny Vega

EuroPython 2025 — South Hall 2B on 2025-07-17] *

An Introduction to Mechanistic Interpretability – Neel Nanda | IASEAI 2025

An Introduction to Mechanistic Interpretability – Neel Nanda | IASEAI 2025

How can we reverse engineer what a neural network is doing? In this IASEAI '25 session, An

Sponsored
Mechanistic Interpretability and How LLMs Understand

Mechanistic Interpretability and How LLMs Understand

A discussion on the philosophy of deep learning,

A Hackers' Guide to Language Models

A Hackers' Guide to Language Models

In this deeply informative video, Jeremy Howard, co-founder of fast.ai and creator of the ULMFiT approach on which all modern ...

Understanding and improving LLMs through mechanistic interpretability

Understanding and improving LLMs through mechanistic interpretability

ACL SIG-FinTech x TFAI Webinar Series (https://sigfintech.github.io/) Understanding and improving

Sponsored
The Dark Matter of AI [Mechanistic Interpretability]

The Dark Matter of AI [Mechanistic Interpretability]

Take your personal data back with Incogni! Use code WELCHLABS at the link below and get 60% off an annual plan: ...

Mechanistic Interpretability for NLP: One-stop Guide for Everything you Need to Know

Mechanistic Interpretability for NLP: One-stop Guide for Everything you Need to Know

0:00

Understanding and improving LLMs through mechanistic interpretability

Understanding and improving LLMs through mechanistic interpretability

Next I will

Exploration Hacking: LLMs Resisting RL Training

Exploration Hacking: LLMs Resisting RL Training

In this AI Research Roundup episode, Alex discusses the paper: 'Exploration

Tracing the thoughts of a large language model

Tracing the thoughts of a large language model

AI models are trained and not directly programmed, so we don't understand how they do most of the things they do. Our new ...

Prompt Injection Methodology for GenAI Application Pentesting - Greet & Repeat Method

Prompt Injection Methodology for GenAI Application Pentesting - Greet & Repeat Method

A 4 Step Prompt Injection Methodology - Greet and Repeat While tons of courses dive into pentesting GenAI-powered ...

LLMjacking: How hackers steal your AI API keys and stick you with the bill

LLMjacking: How hackers steal your AI API keys and stick you with the bill

Explore the podcast → https://ibm.biz/~sW0ssm7Tk AI tools can turn a team of three developers into a fully functioning company.

SenSys 2026 Talk: Jailbreaking Embodied LLMs via Action-Level Manipulation

SenSys 2026 Talk: Jailbreaking Embodied LLMs via Action-Level Manipulation

Presentation video for our paper “Jailbreaking Embodied

Hacking LLMs Demo and Tutorial (Explore AI Security Vulnerabilities)

Hacking LLMs Demo and Tutorial (Explore AI Security Vulnerabilities)

Big thank you to Cisco for sponsoring this video and sponsoring my trip to Cisco Live Amsterdam. // FREE Ethical