Since legal text is often rather long, conventional transformer-based models are not optimal in this area. Various efficient transformer variants have been proposed . However, none of these models has been pretrained in languages other than English so far. Additionally, models pretrained on legal data are rather rare and do not exist for the Swiss national languages so far. To pretrain an efficient multilingual transformer-based model capable of handling long text input on legal data, many corpora need to be collected.
List of possible data sources: