What AI guidelines does your tech organization have in place?

Posted by FewWatercress4917@reddit | ExperiencedDevs | View on Reddit | 26 comments

Both technical and non-technical people at our startup are in love with LLMs - Cursor, Devin, Lovable, etc. I agree that these bring additional capabilities to people to do stuff faster, but I also can't help but notice a downside: Even the most thoughtful senior engineers will, over time, trust the AI more and stop thinking about everything it is doing. If it works, 95% test coverage and e2e playwright tests pass - then it must be good! A few things I am worried about:

  1. Over time, the codebase will start feeling like it was written by 200 different people (we are a 15 person tech team). The standards for getting code in fall by the wayside as people just accept what cursor/devin do.

  2. Stackoverflow and docs get a lot of deserved criticism, but people had a way to judge junk answers vs answers from people who really knew what they were talking about, canonical sources, etc. This is being lost right now and engineers just accept what the AI tells them.

I think these tools bring benefit - but I am starting to be afraid of the downsides (ie, making everyone dumber). How did you address this and how do you use it in your organization?