Future Development
With the publication of the Virtual Light Table in its release version v1.0.0 on 10 August 2023, the development within the Crossing Boundaries project, in the context of which it was conceived and implemented, will end at the same time. However, this does not mean that the Virtual Light Table will not be further developed. So what can be expected for the future?
What's Coming Next?
Below you will find a list of concepts and ideas that are being considered for the further development of the Virtual Light Table. These are mainly major structural changes or far-reaching feature extensions that cannot be done from one day to the next. This list is therefore intended primarily as an indication of the direction in which the Virtual Light Table will be fundamentally developed. However, the mention of an item on this list does not indicate a potential release date. It is also not guaranteed that the features will (be able to) be implemented in this form.
Currently, the VLT only works with bitmap graphics. Therefore, a possible solution could be to simply rasterize the vector graphics and convert them into a readable format this way.
Fortunately, EaselJS, which I use mostly for representing graphics on the VLT, seems to also offer support for SVG graphics (it is to be determined, however, to which extent). It should therefore be technically relatively easy to make this format usable for the VLT as well.
The challenges here lie on the one hand in the data input - is it possible to automatically extract SVG graphics from a PDF or the proprietary Adobe Illustrator format? Or does EaselJS already handle this "out of the box"? And what implications does the use of an SVG graphic have for the further program flow, for example with regard to zooming, masking, graphic filters, the calculation of distances and sizes, etc.?
The results of this research will eventually also find their way into the Virtual Light Table. In the end, the user should be able to create a corpus of her own objects (or use an existing corpus), calculate the corresponding embeddings for these objects and create a sorting in this way. The user can then select individual objects and have the system automatically display which other fragments have the greatest similarities with regard to the aspects he or she has selected (e.g. colour, texture, handwriting, text).
It should also be possible to feed encoded ancient Egyptian texts (in Manual de Codage or Unicode) into the VLT and ask the machine for an assessment according to the text genre. It should also be possible to consult the machine about the potential content of blanks or the continuation of a broken line of text.
Disclaimer: As extraordinary as this all sounds, machine learning applications are not a modern form of magic but calculate probabilities and deductions based on the material available to them during training. The results of such a calculation can serve as an indication of what MAY be true. However, they do not produce absolute truths. Critical interpretation and scrutiny of the results by the trained eye therefore remains unavoidable.
However, there are definitely disadvantages to this. The exchange of research with colleagues is much easier, all data is collected on one server and only the access rights have to be changed. It therefore appears to be a sensible concept for the future to plan and implement the VLT as a web application in addition to its standalone application (which will continue to exist). Of course, special challenges lie here in the provision, financing and maintenance of the necessary infrastructure.
Even though the machines are not trained in this case, but only used for inference, a whole series of calculations take place here, depending on the size and complexity of the model. The larger the input and the more individual calculation steps required, the longer it can take for a calculation to be completed. The speed of the underlying computer configuration also has a significant influence on this.
With the idea of turning the VLT into a web service comes the simultaneous idea of moving the machine learning operations to a server on the network. This can be done either within the framework of existing frameworks such as huggingface. Or actually as a server of its own, which would also offer researchers the possibility to participate voluntarily (via opt-in) in the improvement of the models with their material.
Thus, the creation of annotated data sets is often one of the necessary, albeit time-consuming, components for building useful machine learning models. For example, if one wants to teach a machine to automatically segment not only papyri but also ostraca, this first requires precise information for a number of example images as to which pixels should belong to which pixel class (e.g., ostracon; ink; background) for the machine.
Comfortable tools already exist for the generation of such annotations. Nevertheless, the VLT should also receive corresponding tools in the long run. This will allow users to create their own datasets, to make them available for ML research or just to enhance their images with supplementary information and annotations.
I Have Some Ideas What Should Come Next!
Ideally, an application should be oriented to the needs and wishes of the users. In this respect, I am not only willing, but would also like as many users as possible to come to me with their ideas on how the Virtual Light Table could be made even more useful. The broader the idea, i.e., the larger the potential user group, the more useful it is likely to be for implementation. Suggestions and wishes can therefore always be sent to me.