Use robots.txt to block crawlers from "action URLs." This prevents wasted server resources from useless crawler hits. It's an age-old best practice that remains relevant today. Google's Gary Illyes ...