I don’t think we really observed such artifacts, or at least not due to LPIPS. We did see something similar without LPIPS when using JIT on undertrained models, but I don’t think we observe them anymore. It might simply be because the model had not converged yet.
One situation where we started to observe checkerboard patterns was during reward alignment with GRPO or SRPO on one of these models. In that case the artifacts were much more noticeable, but I don’t think they were related to LPIPS since we were not using it at the time.
What type of model were you using?