{"id":7314,"date":"2025-04-01T19:26:12","date_gmt":"2025-04-01T17:26:12","guid":{"rendered":"https:\/\/it-u.at\/?post_type=professor&#038;p=7314"},"modified":"2026-02-06T20:31:28","modified_gmt":"2026-02-06T19:31:28","slug":"peter-balazs","status":"publish","type":"professor","link":"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/","title":{"rendered":"Peter Balazs"},"content":{"rendered":"\n<h2 class=\"wp-block-heading has-h-4-font-size\">Role at IT:U:<\/h2>\n\n\n\n<p>As a Fellow Professor of <a href=\"https:\/\/it-u.at\/en\/research\/research-groups\/research-group-acoustics-analysis-ai-3ai\/\">Acoustics, Analysis and AI (3AI)<\/a> at IT:U, <a href=\"https:\/\/it-u.at\/en\/research\/professors\/peter-balazs\/\">Peter Balazs<\/a> is actively engaged in research at the IT:U Sound and Acoustics Lab and collaborates with the ARI Lab at the \u00d6AW. His work focuses on harmonic analysis, signal processing, and machine learning, contributing to advanced models in acoustics.\u202f&nbsp;<\/p>\n\n\n\n<h2 class=\"wp-block-heading has-h-4-font-size\">Research group:<\/h2>\n\n\n\n<p>The <a href=\"https:\/\/it-u.at\/en\/research\/research-groups\/research-group-acoustics-analysis-ai-3ai\/\">Acoustics, Analysis &amp; AI (3AI)<\/a> group at the IT:U Austria in Linz explores how mathematics, acoustics, and artificial intelligence can work together. We study interpretable, hybrid AI systems for sound analysis\u2014combining deep learning with mathematical structure and human insight. Our research ranges from signal processing and hearing models to bioacoustics and mathematical foundations, aiming to bridge human and machine understanding of acoustic phenomena.&nbsp;<\/p>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\" style=\"margin-top:var(--wp--preset--spacing--3);margin-bottom:var(--wp--preset--spacing--3)\"\/>\n\n\n\n<div class=\" alignnone alignnone wp-block-atom-research-teaser\">\n    <div class=\"container container-wider\">\n        <div class=\"row\">\n            <div class=\"col-12\">\n                    <p class=\"text-center m-0\">Keine Forschungsgebiete gefunden.<\/p>\n                <\/div>        <\/div>\n    <\/div>\n<\/div>\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\" style=\"margin-top:var(--wp--preset--spacing--3);margin-bottom:var(--wp--preset--spacing--3)\"\/>\n\n\n\n<h2 class=\"wp-block-heading has-h-4-font-size\">Short bio:<\/h2>\n\n\n\n<p><a href=\"https:\/\/it-u.at\/en\/research\/professors\/peter-balazs\/\">Peter Balazs<\/a> is the Director of the Acoustics Research Institute at the \u00d6AW. He possesses extensive expertise in acoustics, signal processing, and mathematical methods. His research has led to significant contributions to frame theory and time-frequency analysis, with applications in numerical acoustics, mathematical physics, psychoacoustics and bioacoustics.&nbsp;&nbsp;<\/p>\n\n\n\n<p>His vision is bridging (deep theoretical) mathematics to more applied scientific topics. He has long been passionately committed to application-oriented mathematics for acoustics, signal-processing and machine learning. This commitment was already evident in 2011, when his project Frames and Linear Operators for Acoustical Modeling and Parameter Estimation was awarded the prestigious START Prize.&nbsp;&nbsp;<\/p>\n\n\n\n<p>His&nbsp;interdisciplinary work integrates mathematical models with acoustic applications, as demonstrated in current projects such as Frames and Time-Frequency Analysis in Machine Learning or Decoding Elephant Communication with AI.&nbsp;&nbsp;<\/p>\n\n\n\n<p>His extensive experience in teaching and supervision, his ability to present complex theoretical concepts in an accessible applied way significantly enriches the academic environment at IT:U. His affiliation with the Austrian Academy of Sciences and his active involvement in international scientific organization \u2013 he currently serves the treasurer of the International Commission for Acoustics, e.g. &#8211; further&nbsp;strengthen research and mentorship at IT:U.\u202f&nbsp;<\/p>\n\n\n\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p class=\"has-copy-small-font-size\">&#8222;<em>My vision is bridging (deep theoretical) mathematics to more applied scientific topics. I am passionately committed to application-oriented mathematics for acoustics, signal-processing and machine learning. This allows the full usage of mathematical rigor and controllability of parameters for models and methods in the applied sciences on one hand. On the other hand, this creates new mathematical topics and concepts, raises novel questions within mathematics that are interesting and inspiring per-se&#8220;.<\/em><\/p>\n<cite>Peter Balazs<\/cite><\/blockquote>\n\n\n\n<div class=\"col-12 col-md-6 col-lg-6 align wp-block-atom-box\">\n\n            <div class=\"box has-white-background-color\">\n                \n        <div class=\"is-layout-constrained\">\n\n<h3 class=\"wp-block-heading\"><strong><strong>IT:U PhD Projects Acoustics, Analysis and AI (3AI)<\/strong><\/strong><\/h3>\n\n\n\n<p>Here we describe six PhD projects envisioned for the group 3AI of fellow professor P. Balazs.<\/p>\n\n\n\n<p>The PhD curriculum at IT:U follows a 4-year path. In the first year, emphasis is placed on focused research lab modules.<\/p>\n\n\n\n<p>The first year concludes with the PhD Proposal Presentation. Over the next three years, the content of the PhD thesis will be developed. This is accompanied by interdisciplinary research seminars and work as a project assistant. The PhD program concludes after the 4th year with the submission and defence of the PhD thesis.<\/p>\n\n\n\n<p>A call will be sent out for three positions, where candidates can apply for any of the projects. (This means that for some projects we will have no candidate.)<\/p>\n\n\n\n<details class=\"wp-block-details is-layout-flow wp-block-details-is-layout-flow\"><summary>Building Improved USV Boxes&nbsp;<\/summary>\n<p>Home-cage monitoring of mice and other laboratory animals are increasingly being used in translational research, but several technical changes are needed in conventional cages to be able to study acoustic communication. In this project we will design and construct a new type of mouse cage with integrated automated control to monitor, record, and playback sonic and ultrasonic vocalisations. The cage will feature integrated video recording alongside an improved acoustic environment, computer-controlled doors, and a microphone\u2013speaker array capable of recording and playing back ultrasonic vocalizations with high fidelity. The video data will complement the microphone array, enabling simultaneous behavioral analysis, enhancing audio source separation, supporting advanced noise reduction, and allowing precise detection of individual vocalizations. Alongside the hardware development, the project will implement signal processing and machine learning algorithms for real-time detection, classification, and synthesis of mouse vocalizations, enabling closed-loop behavioral experiments. Special emphasis will be placed on co-design: optimizing the acoustic environment, minimizing handling of the animals, and ensuring robust, low-latency interaction between software and hardware. The resulting platform will support reproducible, high-throughput studies of communication, learning, and social behavior in laboratory mice.<\/p>\n\n\n\n<p>This project will be conducted together with <a href=\"https:\/\/www.vetmeduni.ac.at\/klivv\/forschung\/verhaltensbiologie-penn-zala-lab\">D. Penn and S. Zala<\/a> (VetMed Vienna). . It will be supported by PostDocs at ARI (<a href=\"https:\/\/www.oeaw.ac.at\/en\/ari\/team\/our-team\/abbasi-reyhaneh\">R. Abbasi)<\/a>.<\/p>\n\n\n\n<p>It will be conducted together with the IT:U HANS \u2013 Digital Production Lab and\/or the IT:U Design Lab.<\/p>\n\n\n\n<p>See e.g.<\/p>\n\n\n\n<p>* S. M. Zala, D. Reitschmidt, A. Noll, P. Balazs, D. J. Penn, &#8222;Sex-Dependent<\/p>\n\n\n\n<p>Modulation of Ultrasonic Vocalizations in House Mice (Mus musculus musculus)&#8220;, PLOS ONE(2017)<\/p>\n<\/details>\n\n\n\n<details class=\"wp-block-details is-layout-flow wp-block-details-is-layout-flow\"><summary>Optimizing Artificial Neural Networks to Probe the Functional Role of Spatial Hearing<\/summary>\n<p>Why is spatial hearing important? Is its primary role to help us localize objects in space, or rather to segregate a sound of interest from competing sources\u2014such as during a cocktail party? Artificial neural networks (ANNs) offer a powerful framework to address such \u201cwhy\u201d questions about brain function. If an ANN trained on a particular task spontaneously reproduces a phenomenon observed in humans\u2014while training on other tasks does not\u2014this suggests that the phenomenon may emerge from optimizing the brain for that same task.<\/p>\n\n\n\n<p>A striking example in spatial hearing is the history-dependent adaptation of absolute space perception. This has been speculated to reflect an adaptation that supports source segregation rather than precise localization.<\/p>\n\n\n\n<p>This project will test that hypothesis by training ANNs on different objectives:<\/p>\n\n\n\n<ol>\n<li>sound source segregation (\u201cwhat?\u201d),<\/li>\n\n\n\n<li>spatial localization (\u201cwhere?\u201d), or<\/li>\n\n\n\n<li>both combined.<\/li>\n<\/ol>\n\n\n\n<p>We will then examine under which training conditions human-like perceptual phenomena emerge. In addition, we will investigate the role of ecological validity by comparing networks trained on naturalistic, reverberant multi-source environments with those trained in simplified, anechoic conditions.<\/p>\n\n\n\n<p>Ultimately, the project will shed light on whether and how representations of what and where interact in spatial hearing, offering insights into the functional relevance of this fundamental ability.<\/p>\n\n\n\n<p>This project will be conducted together with <a href=\"https:\/\/www.oeaw.ac.at\/en\/ari\/team\/our-team\/baumgartner-robert\">R. Baumgartner<\/a> (ARI Vienna). <\/p>\n\n\n\n<p>See e.g.<\/p>\n\n\n\n<p>* Meijer, D., Barumerli, R. &amp; Baumgartner, R. How relevant is the prior? Bayesian causal inference for dynamic perception in volatile environments. eLife 14, (2025).<\/p>\n\n\n\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"800\" height=\"422\" src=\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/headsfront-5.png\" alt=\"\" class=\"wp-image-16895\" srcset=\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/headsfront-5.png 800w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/headsfront-5-300x158.png 300w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/headsfront-5-768x405.png 768w\" sizes=\"(max-width: 800px) 100vw, 800px\" \/><\/figure>\n\n<\/details>\n\n\n\n<details class=\"wp-block-details is-layout-flow wp-block-details-is-layout-flow\"><summary>Machine Learning and Signal Processing for Bioacoustic Vocalizations<\/summary>\n<p>This project aims to develop signal-processing and machine learning tools for automated segmentation, classification, and parameter estimation of animal vocalizations. Building on recent advances in computational bioacoustics, spectrogram- or waveform-based feature extraction, combined with convolutional, recurrent neural network architectures (and possibly other ML approaches) will be used to detect and classify call types, and extract acoustic parameters such as fundamental frequency contours, harmonics, or temporal structure under real-world<\/p>\n\n\n\n<p>conditions with varying noise and domain shifts. Special attention will be paid to generalization: training and evaluating models across different recording conditions and species, designing effective data augmentation schemes, and deploying pretrained models and embeddings. The software and analysis tools will be open- source and interoperable, allowing reproducibility and seamless integration with diverse experimental platforms. The overall goal is to develop of a robust and interpretable framework that enables biologists to scale up quantitative bioacoustic analysis.<\/p>\n\n\n\n<p>This project will be conducted on mice, budgerigars, elephants and potentially other animals, together with <a href=\"https:\/\/www.oeaw.ac.at\/en\/ari\/team\/our-team\/hoeschele-marisa\">M. H\u00f6schele<\/a>, <a href=\"https:\/\/www.oeaw.ac.at\/en\/ari\/team\/our-team\/stoeger-angela\">A. St\u00f6ger<\/a> and\/or <a href=\"https:\/\/www.oeaw.ac.at\/en\/ari\/team\/our-team\/baotic-anton\">A. Baotic<\/a> (ARI Vienna) depending on the animal taxa to be worked on. It will be supported by PostDocs at ARI (<a href=\"https:\/\/www.oeaw.ac.at\/en\/ari\/team\/our-team\/abbasi-reyhaneh\">R. Abbasi<\/a>, <a href=\"https:\/\/www.oeaw.ac.at\/en\/ari\/team\/our-team\/haider-daniel\">D. Haider)<\/a>.<\/p>\n\n\n\n<p>See e.g.<\/p>\n\n\n\n<p>* R. Abbasi, P. Balazs, A. Marconi, D. Nicolakis, S. Zala, D. Penn, &#8222;Capturing the songs of mice with an improved detection and classification method for ultrasonic vocalizations (BootSnap)&#8220;, PLOS Computational Biology, Vol. 18( 5): e1010049, preprint on biorxiv (2022)<\/p>\n<\/details>\n\n\n\n<details class=\"wp-block-details is-layout-flow wp-block-details-is-layout-flow\"><summary>Building Improved Bioacoustic Monitors<\/summary>\n<p><em>This project addresses <\/em>the joint challenge of building energy-efficient acoustic monitoring hardware and <em>developing <\/em>embedded machine learning algorithms <em>is addressed <\/em>for wildlife applications. <em>Specifically, the goals are to <\/em>design and prototype low-power acoustic monitors \u2014 either stationary sensors for zoos and field sites, or collar-mounted devices for large mammals such as giraffes or elephants \u2014 and <em>equip <\/em>them with <em>advanced <\/em>on-device signal processing and learning capabilities. <em>The <\/em>research will mainly focus on co-designing the hardware and algorithms: <em>optimizing microphones, developing efficient power supply solutions, designing data transmission pipelines for long-term deployment, and implementing lightweight machine learning models <\/em>for event detection, call classification, and <em>active <\/em>learning directly on the device. The <em>overall <\/em>goal is a <em>comprehensive <\/em>system that enables autonomous bioacoustic monitoring and real-time tracking <em>with <\/em>minimal energy consumption.<\/p>\n\n\n\n<p><em>The developed system will be used for <\/em>elephants, giraffes, cheetahs, gibbons, or general wildlife monitoring, e.g. in Austrian woods, together with <a href=\"https:\/\/www.oeaw.ac.at\/en\/ari\/team\/our-team\/stoeger-angela\">A. St\u00f6ger<\/a> and\/or <a href=\"https:\/\/www.oeaw.ac.at\/en\/ari\/team\/our-team\/baotic-anton\">A.<\/a> <a href=\"https:\/\/www.oeaw.ac.at\/en\/ari\/team\/our-team\/baotic-anton\">Baotic<\/a> (ARI Vienna) depending on the animal taxa.. It will be supported by PostDocs at ARI (<a href=\"https:\/\/www.oeaw.ac.at\/en\/ari\/team\/our-team\/haider-daniel\">D. Haider<\/a>), and a PhD student at ARI (<a href=\"https:\/\/www.oeaw.ac.at\/en\/ari\/team\/our-team\/zeleznik-jure\">J.Zeleznik<\/a>).<\/p>\n\n\n\n<p>It will be conducted together with the IT:U HANS \u2013 Digital Production Lab and\/or the IT:U Design Lab.<\/p>\n\n\n\n<p>See e.g.<\/p>\n\n\n\n<p>* Zeppelzauer, M., Hensman, S., &amp; Stoeger, A. S. (2014). Towards an automated<\/p>\n\n\n\n<p>acoustic detection system for free-ranging elephants. <em>Bioacoustics<\/em>, <em>24<\/em>(1), 13\u201329. https:\/\/doi.org\/10.1080\/09524622.2014.906321<\/p>\n<\/details>\n\n\n\n<details class=\"wp-block-details is-layout-flow wp-block-details-is-layout-flow\"><summary>An Interpretable Hearing Model<\/summary>\n<p>We will develop an interpretable (human) hearing model, building on our research on stable neural networks, and aimed at a continuous learning approach predicting human behavior in psychoacoustic experiments. In contrast to existing concepts, for the first task, we will create an interpretable invertible neural network-approach with learned filterbanks. The pipeline of a first model could be a hybrid filterbank based on an AUDlet, followed by a temporal convolution network (tCN) to model temporal integration in hearing, and general INNs, with a task-specific decision network at the end. The level of complexity will be increased by including tensors, recursive, and concatenated filterbank approaches to incorporate published modeling attempts such as multiple looks, level dependency and dual pathway approaches. We will learn and test our methods with a transfer learning approach on existing psychoacoustic datasets, predicting the results in published papers and integrate the results in the auditory modeling toolbox, hosted at ARI.<\/p>\n\n\n\n<p>This project will be conducted together with <a href=\"https:\/\/www.oeaw.ac.at\/en\/ari\/team\/our-team\/laback-bernhard\">B. Laback<\/a> (ARI). See e.g.<\/p>\n\n\n\n<p>Saddler MR, McDermott JH. Models optimized for real-world tasks reveal the task- dependent necessity of precise temporal coding in hearing. Nat Commun. 2024 Dec 4;15(1):10590. doi: 10.1038\/s41467-024-54700-5. PMID: 39632854; PMCID:<\/p>\n\n\n\n<p>PMC11618365. <a href=\"https:\/\/www.nature.com\/articles\/s41467-024-54700-5\">https:\/\/www.nature.com\/articles\/s41467-024-54700-5<\/a><\/p>\n<\/details>\n\n\n\n<details class=\"wp-block-details is-layout-flow wp-block-details-is-layout-flow\"><summary>Interpretable Audio-Preprocessing<\/summary>\n<p>An ongoing debate in the design of machine learning models, particularly for applications involving acoustic data, is how input signals should be pre-processed and analyzed. <em>That is, what is the best approach to feature extraction for audio? <\/em>The most common approach uses fixed time-frequency representations <em>as a human- engineered way of representing audio to the model<\/em>. An alternative approach is to \u201clet the machine learn <em>the representation<\/em>\u201d by using the raw audio signal directly as model input. <em>The aim of this project is to explore the connections between concepts from signal processing and machine learning from a theoretical and practical point of view, and develop explanations as well as novel designs of feature extractors that leverage knowledge from both fields. An existing line of work within the project concerns a \u201chybrid\u201d approach that combines classical filterbank design with an integrated learning step by the model. A future goal here is to embed this construction into auditory research and design and use learned modulation filterbanks for feature extraction. <\/em>Additional topics include the interpretation of recurrent convolutional layers as filterbanks on infinite sequence spaces and the link of gradient boosting to empirical mode decomposition (EMD).<\/p>\n\n\n\n<p>This project will be conducted together with <a href=\"https:\/\/www.lostanlen.com\/\">V. Lostanlen<\/a> (LS2N CNRS Nantes) and will be supported by PostDocs at ARI (<a href=\"https:\/\/www.oeaw.ac.at\/en\/ari\/team\/our-team\/haider-daniel\">D. Haider<\/a>).<\/p>\n\n\n\n<p>See e.g.<\/p>\n\n\n\n<p>* D. Haider\u2217 , F. Perfler\u2217 , V. Lostanlen, M. Ehler, P. Balazs, &#8222;Hold Me Tight: Stable Encoder\u2013Decoder Design for Speech Enhancement&#8220;, Interspeech 2024, Kos Island (2024)<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"304\" src=\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/filters-1024x304.png\" alt=\"\" class=\"wp-image-16899\" srcset=\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/filters-1024x304.png 1024w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/filters-300x89.png 300w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/filters-768x228.png 768w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/filters-1536x456.png 1536w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/filters.png 1787w\" sizes=\"(max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n<\/details>\n\n<\/div>\n\n        \n            <\/div> \n        \n<\/div>\n\n\n<div class=\"col-12 col-md-6 col-lg-6 align wp-block-atom-box\">\n\n            <div class=\"box has-white-background-color\">\n                \n        <div class=\"is-layout-constrained\">\n\n<h3 class=\"wp-block-heading\"><strong><strong>Potential Projects<\/strong><\/strong><\/h3>\n\n\n\n<details class=\"wp-block-details is-layout-flow wp-block-details-is-layout-flow\"><summary><strong>Potential Master and Off-Lab Projects:<\/strong>&nbsp;<br>&nbsp;<\/summary>\n<ul>\n<li>Building a 3D Audio Recorder from scratch&nbsp;<\/li>\n<\/ul>\n\n\n\n<ul>\n<li>Building a 3D Audio Sound system from scratch&nbsp;<\/li>\n<\/ul>\n\n\n\n<ul>\n<li>Building a Measurement System for HRTFs from scratch&nbsp;<\/li>\n<\/ul>\n\n\n\n<ul>\n<li>Improve 3D Audio in other IT-U labs&nbsp;<\/li>\n<\/ul>\n\n\n\n<ul>\n<li>Build an Audiometry (from scratch)&nbsp;<\/li>\n<\/ul>\n\n\n\n<ul>\n<li>Do a sub-project of PhD projects focusing on certain details, e.g.&nbsp;&nbsp;<\/li>\n<\/ul>\n<\/details>\n\n\n\n<details class=\"wp-block-details is-layout-flow wp-block-details-is-layout-flow\"><summary><strong>Development of bioacoustics and movement monitoring collar mounted sensors for African Savannah elephants (<em>Loxodonta Africana<\/em>) \u2013 MSc thesis project<\/strong>&nbsp;<\/summary>\n<p><em>Supervision: Prof. Dr. Peter Balazs\u202f &amp; Prof. Dr. Angela St\u00f6ger<\/em>&nbsp;<br><em>Co-supervision: Dr. Daniel Haider &amp; Jure \u017deleznik<\/em>&nbsp;<br>&nbsp;<\/p>\n\n\n\n<p>Collar-mounted wildlife sensors enable continuous, minimally invasive monitoring of behavior,\u202fhabitat use, and inter-individual acoustic communication. Yet, attributing specific vocalizations to the correct individual in free-ranging groups remains challenging when audio is collected without tightly synchronized motion and position data. The candidate in this project will design and prototype a field-ready elephant collar that integrates wide-band acoustic recording, high-rate tri-axial accelerometery (X\/Y\/Z), and GPS positioning in a time-synchronized, low-power package. The system will feature precise clocking and sensor fusion pipelines so that subtle body movements and collar vibrations can be aligned with acoustic onsets, while GPS trajectories provide spatial constraints, together enabling robust focal-caller attribution. Emphasis will be placed on the ability to detect focal callers as well as power budgeting for longer deployments, rugged enclosure and microphone placement to minimize environmental noise, and serviceable data logging. The outcome will be a validated hardware prototype with open documentation, calibration and synchronization procedures, and a thesis evaluating focal-caller detection accuracy from combined sensors versus audio-only baselines.&nbsp;<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"609\" src=\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/adobestock_85248976-1024x609.jpeg\" alt=\"\" class=\"wp-image-16908\" srcset=\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/adobestock_85248976-1024x609.jpeg 1024w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/adobestock_85248976-300x178.jpeg 300w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/adobestock_85248976-768x457.jpeg 768w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/adobestock_85248976-1536x913.jpeg 1536w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/adobestock_85248976-2048x1218.jpeg 2048w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/adobestock_85248976-1800x1070.jpeg 1800w\" sizes=\"(max-width: 1024px) 100vw, 1024px\" \/><span class=\"credits\">&copy; AdobeStock<\/span><\/figure>\n\n<\/details>\n\n\n\n<details class=\"wp-block-details is-layout-flow wp-block-details-is-layout-flow\"><summary><strong>Development of sound event annotating tool with incorporated few-shot and active learning \u2013 MSc thesis project<\/strong><\/summary>\n<p><em>Supervision: Prof. Dr. Peter Balazs\u202f &amp; Prof. Dr. Angela St\u00f6ger<\/em>&nbsp;<br><em>Co-supervision: Dr. Daniel Haider &amp; Jure \u017deleznik<\/em>&nbsp;<br>&nbsp;<\/p>\n\n\n\n<p>The use of long-term autonomous recorders has been steadily increasing over the past few decades in the field of bioacoustics. This means that wildlife researchers and bioacousticians are collecting thousands of hours of recordings, after which they manually extract sound events of their interest. Although the development and use of machine learning algorithms for automatic sound event extraction and analysis are promising, these generally only work well on large well-defined curated datasets with previously identified sounds events that can be used for training and validating. To make this available to the dataset at hand, the candidate in this project will design and prototype a modern annotation platform for audio and bioacoustics research that can couple an intuitive labelling UI with state-of-the-art machine learning. The tool will provide waveform\/spectrogram views, rapid keyboard\/mouse labelling, and flexible schema management, facilitating few-shot learners to \u201ccold-start\u201d models from just a handful of examples per class. An active-learning loop will continuously surface the most informative or uncertain clips for review, maximizing annotation efficiency and model quality with minimal human effort.\u202fEmphasis will be placed on the development of a tool which can be used for wildlife monitoring and\/or environmental soundscape analysis. The outcome will be a usable, extensible open-source tool, and a thesis evaluating gains in labelling speed\/accuracy versus conventional workflows.<\/p>\n<\/details>\n\n<\/div>\n\n        \n            <\/div> \n        \n<\/div>\n\n\n<div class=\"position-relative alignfull wp-block-atom-slider-wrapper\">\n    <div class=\"slick-slider\">\n\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"800\" height=\"422\" src=\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/headsfront-5.png\" alt=\"\" class=\"wp-image-16895\" srcset=\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/headsfront-5.png 800w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/headsfront-5-300x158.png 300w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/headsfront-5-768x405.png 768w\" sizes=\"(max-width: 800px) 100vw, 800px\" \/><\/figure>\n\n\n\n\n<figure class=\"wp-block-image size-full is-resized\"><img loading=\"lazy\" decoding=\"async\" width=\"1921\" height=\"963\" src=\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/jumphann8.png\" alt=\"\" class=\"wp-image-16897\" style=\"width:1450px;height:auto\" srcset=\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/jumphann8.png 1921w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/jumphann8-300x150.png 300w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/jumphann8-1024x513.png 1024w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/jumphann8-768x385.png 768w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/jumphann8-1536x770.png 1536w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/jumphann8-1800x902.png 1800w\" sizes=\"(max-width: 1921px) 100vw, 1921px\" \/><\/figure>\n\n\n\n\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"1921\" height=\"939\" src=\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/synth.png\" alt=\"\" class=\"wp-image-16901\" srcset=\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/synth.png 1921w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/synth-300x147.png 300w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/synth-1024x501.png 1024w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/synth-768x375.png 768w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/synth-1536x751.png 1536w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/synth-1800x880.png 1800w\" sizes=\"(max-width: 1921px) 100vw, 1921px\" \/><\/figure>\n\n\n\n\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"1974\" height=\"1536\" src=\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/frame2.png\" alt=\"\" class=\"wp-image-16903\" srcset=\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/frame2.png 1974w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/frame2-300x233.png 300w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/frame2-1024x797.png 1024w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/frame2-768x598.png 768w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/frame2-1536x1195.png 1536w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/frame2-1800x1401.png 1800w\" sizes=\"(max-width: 1974px) 100vw, 1974px\" \/><\/figure>\n\n\n\n\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"1787\" height=\"531\" src=\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/filters.png\" alt=\"\" class=\"wp-image-16899\" srcset=\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/filters.png 1787w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/filters-300x89.png 300w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/filters-1024x304.png 1024w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/filters-768x228.png 768w, https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/filters-1536x456.png 1536w\" sizes=\"(max-width: 1787px) 100vw, 1787px\" \/><\/figure>\n\n\n<\/div>\n\n    <\/div>\n\n","protected":false},"template":"","meta":{"_acf_changed":false},"acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v22.5 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Peter Balazs | IT:U<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/\" \/>\n<meta property=\"og:locale\" content=\"de_DE\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Peter Balazs | IT:U\" \/>\n<meta property=\"og:description\" content=\"Role at IT:U: As a Fellow Professor of Acoustics, Analysis and AI (3AI) at IT:U, Peter Balazs is actively engaged in research at the IT:U Sound and Acoustics Lab and collaborates with the ARI Lab at the \u00d6AW. His work focuses on harmonic analysis, signal processing, and machine learning, contributing to advanced models in acoustics.\u202f&nbsp; [&hellip;]\" \/>\n<meta property=\"og:url\" content=\"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/\" \/>\n<meta property=\"og:site_name\" content=\"IT:U\" \/>\n<meta property=\"article:modified_time\" content=\"2026-02-06T19:31:28+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/headsfront-5.png\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Gesch\u00e4tzte Lesezeit\" \/>\n\t<meta name=\"twitter:data1\" content=\"12\u00a0Minuten\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"WebPage\",\"@id\":\"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/\",\"url\":\"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/\",\"name\":\"Peter Balazs | IT:U\",\"isPartOf\":{\"@id\":\"https:\/\/it-u.at\/de\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/headsfront-5.png\",\"datePublished\":\"2025-04-01T17:26:12+00:00\",\"dateModified\":\"2026-02-06T19:31:28+00:00\",\"breadcrumb\":{\"@id\":\"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/#breadcrumb\"},\"inLanguage\":\"de\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"de\",\"@id\":\"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/#primaryimage\",\"url\":\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/headsfront-5.png\",\"contentUrl\":\"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/headsfront-5.png\"},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"home\",\"item\":\"https:\/\/it-u.at\/de\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Peter Balazs\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/it-u.at\/de\/#website\",\"url\":\"https:\/\/it-u.at\/de\/\",\"name\":\"it:u\",\"description\":\"\",\"publisher\":{\"@id\":\"https:\/\/it-u.at\/de\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/it-u.at\/de\/?s={search_term_string}\"},\"query-input\":\"required name=search_term_string\"}],\"inLanguage\":\"de\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/it-u.at\/de\/#organization\",\"name\":\"it:u - interdisciplinary transformation university austria\",\"url\":\"https:\/\/it-u.at\/de\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"de\",\"@id\":\"https:\/\/it-u.at\/de\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/it-u.at\/wp-content\/uploads\/2024\/09\/06\/itu-org-logo.jpg\",\"contentUrl\":\"https:\/\/it-u.at\/wp-content\/uploads\/2024\/09\/06\/itu-org-logo.jpg\",\"width\":696,\"height\":696,\"caption\":\"it:u - interdisciplinary transformation university austria\"},\"image\":{\"@id\":\"https:\/\/it-u.at\/de\/#\/schema\/logo\/image\/\"}}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Peter Balazs | IT:U","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/","og_locale":"de_DE","og_type":"article","og_title":"Peter Balazs | IT:U","og_description":"Role at IT:U: As a Fellow Professor of Acoustics, Analysis and AI (3AI) at IT:U, Peter Balazs is actively engaged in research at the IT:U Sound and Acoustics Lab and collaborates with the ARI Lab at the \u00d6AW. His work focuses on harmonic analysis, signal processing, and machine learning, contributing to advanced models in acoustics.\u202f&nbsp; [&hellip;]","og_url":"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/","og_site_name":"IT:U","article_modified_time":"2026-02-06T19:31:28+00:00","og_image":[{"url":"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/headsfront-5.png"}],"twitter_card":"summary_large_image","twitter_misc":{"Gesch\u00e4tzte Lesezeit":"12\u00a0Minuten"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"WebPage","@id":"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/","url":"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/","name":"Peter Balazs | IT:U","isPartOf":{"@id":"https:\/\/it-u.at\/de\/#website"},"primaryImageOfPage":{"@id":"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/#primaryimage"},"image":{"@id":"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/#primaryimage"},"thumbnailUrl":"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/headsfront-5.png","datePublished":"2025-04-01T17:26:12+00:00","dateModified":"2026-02-06T19:31:28+00:00","breadcrumb":{"@id":"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/#breadcrumb"},"inLanguage":"de","potentialAction":[{"@type":"ReadAction","target":["https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/"]}]},{"@type":"ImageObject","inLanguage":"de","@id":"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/#primaryimage","url":"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/headsfront-5.png","contentUrl":"https:\/\/it-u.at\/wp-content\/uploads\/2026\/02\/06\/headsfront-5.png"},{"@type":"BreadcrumbList","@id":"https:\/\/it-u.at\/de\/research\/professors\/peter-balazs\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"home","item":"https:\/\/it-u.at\/de\/"},{"@type":"ListItem","position":2,"name":"Peter Balazs"}]},{"@type":"WebSite","@id":"https:\/\/it-u.at\/de\/#website","url":"https:\/\/it-u.at\/de\/","name":"it:u","description":"","publisher":{"@id":"https:\/\/it-u.at\/de\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/it-u.at\/de\/?s={search_term_string}"},"query-input":"required name=search_term_string"}],"inLanguage":"de"},{"@type":"Organization","@id":"https:\/\/it-u.at\/de\/#organization","name":"it:u - interdisciplinary transformation university austria","url":"https:\/\/it-u.at\/de\/","logo":{"@type":"ImageObject","inLanguage":"de","@id":"https:\/\/it-u.at\/de\/#\/schema\/logo\/image\/","url":"https:\/\/it-u.at\/wp-content\/uploads\/2024\/09\/06\/itu-org-logo.jpg","contentUrl":"https:\/\/it-u.at\/wp-content\/uploads\/2024\/09\/06\/itu-org-logo.jpg","width":696,"height":696,"caption":"it:u - interdisciplinary transformation university austria"},"image":{"@id":"https:\/\/it-u.at\/de\/#\/schema\/logo\/image\/"}}]}},"_links":{"self":[{"href":"https:\/\/it-u.at\/de\/wp-json\/wp\/v2\/professor\/7314"}],"collection":[{"href":"https:\/\/it-u.at\/de\/wp-json\/wp\/v2\/professor"}],"about":[{"href":"https:\/\/it-u.at\/de\/wp-json\/wp\/v2\/types\/professor"}],"wp:attachment":[{"href":"https:\/\/it-u.at\/de\/wp-json\/wp\/v2\/media?parent=7314"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}