Recap: How do you do it? A behind-the-scenes look at research workflows (2025)
Every academic year, the HDYDI (How Do You Do (It)?) event on research data workflows signals the start of the Digital Scholarship Module. Through a series of sessions and (mini-)workshops, Artes Research aims to guide students through the complexities of scholarship in the digital age, from Open Science to Research Data Management and beyond.
At the HDYDI kick-off event, we invite three researchers from the Faculty of Arts to open the black box of their research workflows. By sharing the practical tools, decisions, and challenges that shape their day‑to‑day work, they aim to offer the first-year PhD researchers a realistic insight into what digital scholarship can look like across disciplines. We hope these behind‑the‑scenes glimpses help you discover approaches that can inform your own research journey!
Tim Debroyer: From Paper to Digital Source
The first speaker, Tim Debroyer, is a third-year PhD candidate at the Cultural History since 1750 research group. Under the supervision of Joris Vandendriessche and Kaat Wils, Tim is studying the evolution of 20th-century Belgian patient organisations as an overlooked link in the development of the modern welfare state. This involves examining their oral history as well as archival and published sources.
The focus of Tim’s talk is on the latter – periodicals specifically form one of the most important sources of information for his project. Faced with thousands of pages early on in his research project, he had to make strategic decisions: what to photograph, how to photograph it, and which digital methods were worth the investment.
Taking BVS Nieuws, the periodical of a diabetes association founded in the 1940s, as an example, Tim explains that he ended up manually photographing the entire series of journals so as to allow for a more thorough discourse analysis. This experience taught him some “tricks” which might be useful to others looking to photograph large amounts of text. Firstly, he used a classic camera in order to avoid the post-processing which smartphones tend to apply, and which can harm OCR quality. Secondly, he made sure to always photograph beyond the edges of the page to make it easier for the OCR software to recognize the boundaries. Thirdly, since taking pictures in the library was quite hectic, Tim always made notes of what he was doing: for instance, what stood out in the issues and what was missing – this made it much easier to return to the sources later on in his trajectory.
Once he properly organized the resulting pictures in folders per issue or volume with short, meaningful names, Tim set to extract the text using OCR (Optical Text Recognition) tools in order to enable keyword searches and quantitative analysis. (This is a labor-intensive step, he cautions, so make sure that it makes sense for your methodology before adopting it yourself.) Numerous scanning apps and online tools exist – Tesseract, Google Cloud Vision and Transkribus (for handwritten text) are great options for the more technically minded – but Tim made use of ABBYY FineReader, a commonly used OCR tool that is very performant and user-friendly. It is a commercial tool, but computers with ABBYY licenses are available at the Maurits Sabbe Library and Agora, so researchers looking to digitize a limited number of sources are free to go there without having to purchase their own license. ABBYY FineReader allows for image pre-processing (e.g. fixing lighting, straightening and cropping pictures), supports various languages, recognizes images in sources as well, and offers various formats for exporting (including .txt files). Tim was quite satisfied with the quality of the OCR’d texts: take good pictures, he says, and ABBYY will deliver good results!
To conclude, Tim shows how he processed the resulting text files in AntConc, a free concordance tool that’s often used for text mining. It allows for large-scale word searching and analysis, can provide keyword frequencies and information about relations to other words, and can easily compare different corpora. (Tim provides a small tip for those looking to explore AntConc: keep a stopword list of high-frequency words with little thematic content that the tool can filter out of its analysis.)
Of course, every researcher has to figure out what workflow suits them, but Tim importantly highlights that you should think about what you want to achieve before investing in digital methods. Consider the nature of your research project, the characteristics of your source corpus, the methodologies you use (discourse analysis, quantitative analysis, network & visual analysis) and let these things decide how you will process and study your sources. At the same time, don’t be afraid to try out new tools that might work well for you!

Of course, the quality of ABBYY FineReader’s OCR results depends on the quality of the input images.
Lauren Ottaviani: Mapping and Analyzing Women’s Magazine Archives
Our second speaker is Lauren Ottaviani, fourth-year PhD candidate in English Literature. Lauren’s project, supervised by Elke D’hoker, focuses on the representation of the women’s suffrage movement in two conservative, middlebrow periodicals dating to the late 19th and early 20th centuries: The Woman at Home and Lady of the House. In doing so, the research seeks to consider the interaction between suffrage and domestic ideals at the turn of the twentieth century.
Similarly to Tim, then, Lauren also works with a large corpus of periodicals; and just as we saw with Tim, many of the magazines’ issues – which tend to be quite lengthy – remained as yet undigitized. The complexity of her materials meant that Lauren had to decide early on how to approach data management efficiently. In the end, a combination of three tools informed her research workflow.
Firstly, early on, she shifted from using Word for note-taking to using the free open-source tool Obsidian instead. As Lauren says, Obsidian (which was covered in last year’s HDYDI session as well) has the same ease of use that a program like Word offers, but you’ll actually be able to find your note again! With its added functionality, Obsidian allowed her to create a relational database of notes categorized by date, theme, or type, so as to keep track of any stories worth revisiting. Through tags and linked notes, Lauren could keep track of authorship, include direct links to the digitized magazine pages, and even uncover recurring anonymous authors. It’s also just a great tool for conference notes and miscellaneous admin.
Secondly, Lauren made use of the storage that’s provided by KU Leuven on OneDrive for Business. Currently, OneDrive is no longer recommended as a primary storage solution for research data at the university,1 but it does have some useful features – and it proved particularly handy for Lauren’s use case. Using the OneDrive smartphone app, she took pictures of interesting articles in the periodicals she was studying and placed those in her pre-organized folder structure. In contrast to Tim, Lauren did not think full OCR of her corpus was worth the time investment or really relevant to her research questions, but this smaller-scale scanning process (which resulted in perfectly legible captures) worked great for her methodology.
Thirdly and finally, Lauren also adopted Nodegoat as part of her workflow, mainly for its “mapping” potential. That is, Nodegoat is a database tool, but it also offers built-in network visualization capabilities, which Lauren used to map out different entries – i.e. letters from the magazines’ correspondence columns – tagged with geolocations. The resulting visualization allowed her to track where readers lived, what the magazines’ geographical reach was, and how their readership expanded over time – elements that were central to her analysis of the periodicals’ circulation.
Using a combination of these three tools, Lauren was able to create a structured, well-organized database out of a vast, undigitized corpus; and even though her approach differed quite substantially from that of Tim, both illustrate how the right tools, used well, help make large-scale periodical research manageable.

Using Nodegoat, Lauren was able to map out the readership of the periodicals she’s studying.
Sinem Bilican: Managing Multimodal Data in Healthcare Research
Sinem Bilican is the last speaker: as a PhD candidate at the Research Unit Translation & Interpreting Studies, she is part of the interdisciplinary research project Managing Language Barriers in Unplanned Care (MaLBUC). With the help of her supervisor Heidi Salaets, Sinem studies linguistic diversity and multilingual communication in healthcare practices with the goal of laying bare overlooked communication barriers. As such, her project involves collaboration with the Faculty of Medicine, and we can reasonably expect very different data types from what we saw in Tim’s and Lauren’s presentations.
Indeed, the interdisciplinary and collaborative nature of the research project – which encompasses ethnographic observations as well as a large-scale survey and interviews – necessitates the implementation of clear research data management practices. Sinem works with extensive field notes, images, video and audio recordings, questionnaires, and other survey data: a lot of materials to manage, to be sure!
Sinem begins by outlining the tools involved in her daily research workflow. Zotero is a usual suspect here, and one which we see in many researchers’ workflows as a handy reference manager as well as a note-taking and annotation tool. OneDrive, meanwhile, enables Sinem to exchange data, drafts and other documents transparently between team members; whereas for a related larger-scale project, the team opted for the ease of use of Teams and SharePoint (which is a recommended storage solution at the Faculty of Arts). Finally, Obsidian is mentioned again, and Sinem stresses its convenience for taking both academic and miscellaneous notes.
Next, Sinem presents some of the tools she used during the data collection phase of her research project. Interestingly, the first tool she talks about is an actual physical tool: a Livescribe pen. This smart pen with a built-in recorder synchronizes handwritten notes with audio, allowing Sinem to easily reconstruct interviews and medical consultations she attended2 – after a day of fieldwork, you can just plug it into your laptop and have everything appear in the Livescribe app. For the surveys, Sinem uses REDCap, which is commonly used in the Biomedical Sciences: it is a highly secure, KU Leuven-authenticated tool that can automatically generate full survey reports. It is, as Sinem points out, also quite a technical tool, but the university provides comprehensive support for users.
The last tool Sinem considers takes us from data collection to research dissemination – namely, Canva. Canva is a user-friendly, web-based design platform that’s great for making posters, visuals, and any other materials you might need to present your research. It allows for image upscaling, QR-code generation, and even themed PowerPoint slide decks. Sinem’s enthusiasm for Canva is infectious – and fittingly, she used it to create her HDYDI presentation as well!
By combining these tools, Sinem is able to navigate a complex, interdisciplinary project that involves varied datasets with clarity and structure; and while her workflow differs markedly from those of Tim and Lauren, it likewise shows how thoughtful tool choices can make even the most challenging research environments manageable.

REDCap proved a useful tool for Sinem’s research data workflow.
Across all three presentations, the workflows we saw revealed both overlaps and differences, but the shared message was clear: the best workflow is the one that genuinely works for your project. Let these examples inspire you, try out the tools that seem useful, and keep what supports your work. With a bit of exploration, you may find a data workflow that not only suits your project, but strengthens it!
- As explained in the university’s storage solution FAQ, there are a number of reasons why OneDrive is no longer recommended as a primary solution for long-term research data storage; most significantly the fact that data stored on OneDrive servers is inaccessible to KU Leuven, which goes against RDM policy (principle II). This means that any data that you’ve kept on OneDrive is erased as soon as you leave the university for any reason, and recovering files is a difficult and costly procedure.
︎ - Of course, these recordings were made with informed consent of all involved.
︎