Section 1 - Instruction

You've learned how AI uses tools and follows system prompts. But what happens when AI says "no" to your study requests?

AI refusals aren't glitches - they're deliberate safety features programmed by developers to maintain academic integrity and prevent harmful outputs.

Engagement Message

Have you ever had an AI refuse to help with your homework or research?

Section 2 - Instruction

AI refuses requests for three main reasons: safety guardrails, technical limitations, and policy compliance.

Safety refusals prevent harmful content like hate speech or dangerous instructions. Technical refusals happen when AI lacks knowledge or capability.

Engagement Message

Which type of refusal would frustrate you more - safety or technical?

Section 3 - Instruction

When AI says "I can't help with that because it may be harmful," that's an ethical refusal based on programmed values.

These aren't the AI's personal moral choices - they're rules set by human programmers and ethicists.

Engagement Message

Should AI have more or fewer restrictions on what it can discuss?

Section 4 - Instruction

Technical refusals sound like "I don't have access to your school's current syllabus" or "I can't check your university's specific grading rubric."

These happen when requests exceed the AI's actual knowledge of your specific academic context or real-time information.

Engagement Message

What's a reasonable technical limitation you'd expect when AI helps with your studies?

Section 5 - Instruction

Sometimes AI gives disclaimers instead of full refusals: "I can't do your homework, but I can explain the concepts to help you learn..."

This acknowledges educational boundaries while still trying to support your learning within appropriate limits.

Engagement Message

Do you prefer AI that explains concepts or one that directly gives you answers?

Section 6 - Instruction
Sign up
Join the 1M+ learners on CodeSignal
Be a part of our community of 1M+ users who develop and demonstrate their skills on CodeSignal