For proprietary AI systems, such as the widely known large language models, training methods and the data used are totally intransparent. Cases have been discussed where personal data was used in the training and could be recovered from the working models. Open source AI tools offer more transparency on some of the development issues, but will this help to respect fundamental rights such as data protection?