Making LLMs safer is more intuitive than you think: How Common Sense and Diversity Improve AI Alignment – BlueDot Impact
Writing Intensive (2024 Dec)

Making LLMs safer is more intuitive than you think: How Common Sense and Diversity Improve AI Alignment

By Jeba Sania (Published on January 28, 2025)

This project was one of the top submissions on the (Dec 2024) Writing Intensive course. The text below is an excerpt from the final project.

AI safety isn't purely technical; it's also about applying common sense and human reasoning. By using reasoning techniques from around the world instead of just the Global North, we can better align AI with human values. If you are interested in AI safety but have an untraditional background or skill set, don't fret. That’s precisely why your ideas are needed.

Full project

View the full project here.

We use analytics cookies to improve our website and measure ad performance. Cookie Policy.