<?xml version="1.0" encoding="utf-8"?>
<rss version="2.0">
	<channel>
		<title><![CDATA[Modartt user forum - Offline version of the Magenta Audio to Midi transcription tool]]></title>
		<link>https://forum.modartt.com/viewtopic.php?id=8999</link>
		<description><![CDATA[The most recent posts in Offline version of the Magenta Audio to Midi transcription tool.]]></description>
		<lastBuildDate>Mon, 20 Dec 2021 02:29:38 +0000</lastBuildDate>
		<generator>PunBB</generator>
		<item>
			<title><![CDATA[Re: Offline version of the Magenta Audio to Midi transcription tool]]></title>
			<link>https://forum.modartt.com/viewtopic.php?pid=979611#p979611</link>
			<description><![CDATA[<div class="quotebox"><cite>Vagporto wrote:</cite><blockquote><div class="quotebox"><cite>budo wrote:</cite><blockquote><p>respectfully i must disagree.&nbsp; or at least i have to say, these arguments are not convincing to me.</p></blockquote></div><p>And I, also respectfully, invite you to read the technical paper by the authors of the method called Onsets and Frames, and you will see why I raised my doubts that I maintain. The deconvolution techniques that are at the core of the method, are not capable of making that distinction, in other than in notes in clear sonic separation of other notes. And even then I believe the separation could only be made by internal comparison, ie by comparing the sonic signature of the same note played in several different moments. </p><p>It is much more possible (although it would amount to the complexity of the best AI programming) to calculate pedaling through the estimation of finger position and movements.</p></blockquote></div><p>no problem, we all have respect here <i class="far fa-smile smiley"></i>.&nbsp; i have read the paper.&nbsp; i&#039;m not advocating doing what they did, or tweaking it, but rather something much more vague: developing a new magenta model.&nbsp; i have no idea if it&#039;s possible, and i&#039;m not an expert, but i also am not convinced that progress couldn&#039;t be made.&nbsp; if one had told me 10 years ago that onsets and frames would exist, i would have been very suprised (actually i would have been just as surprised at the existence of Pianoteq, but i guess that&#039;s another topic).&nbsp; any model would be great, whether based on estimation of finger position or whatever else one thinks is the right way forward.&nbsp; i was just drawn more to a purely sonic model as a first approximation.&nbsp; but i probably will never take it on anyway ... too many other incomplete projects.</p>]]></description>
			<author><![CDATA[null@example.com (budo)]]></author>
			<pubDate>Mon, 20 Dec 2021 02:29:38 +0000</pubDate>
			<guid>https://forum.modartt.com/viewtopic.php?pid=979611#p979611</guid>
		</item>
		<item>
			<title><![CDATA[Re: Offline version of the Magenta Audio to Midi transcription tool]]></title>
			<link>https://forum.modartt.com/viewtopic.php?pid=979610#p979610</link>
			<description><![CDATA[<div class="quotebox"><cite>budo wrote:</cite><blockquote><p>respectfully i must disagree.&nbsp; or at least i have to say, these arguments are not convincing to me.</p></blockquote></div><p>And I, also respectfully, invite you to read the technical paper by the authors of the method called Onsets and Frames, and you will see why I raised my doubts that I maintain. The deconvolution techniques that are at the core of the method, are not capable of making that distinction, in other than in notes in clear sonic separation of other notes. And even then I believe the separation could only be made by internal comparison, ie by comparing the sonic signature of the same note played in several different moments. </p><p>It is much more possible (although it would amount to the complexity of the best AI programming) to calculate pedaling through the estimation of finger position and movements.</p>]]></description>
			<author><![CDATA[null@example.com (Vagporto)]]></author>
			<pubDate>Mon, 20 Dec 2021 00:53:31 +0000</pubDate>
			<guid>https://forum.modartt.com/viewtopic.php?pid=979610#p979610</guid>
		</item>
		<item>
			<title><![CDATA[Re: Offline version of the Magenta Audio to Midi transcription tool]]></title>
			<link>https://forum.modartt.com/viewtopic.php?pid=979605#p979605</link>
			<description><![CDATA[<p>On a very good quality audio file I think that the software could indeed do as you define it, but the interest to process in midi a very good file is limited. To revive a poor quality file as I tried to do with &quot;Duke&quot; the software has great difficulties. And indeed it takes very frequently harmonics for notes, it can of course progress but the difference between a strong harmonic and the note really played softly can be extremely difficult to detect on an old file extremely parasited by other noises. And in this case the resonances with or without pedals are sometimes very blurred. However, reworking the midi file by listening to the original in parallel is very interesting,</p>]]></description>
			<author><![CDATA[null@example.com (YvesTh)]]></author>
			<pubDate>Sun, 19 Dec 2021 23:08:24 +0000</pubDate>
			<guid>https://forum.modartt.com/viewtopic.php?pid=979605#p979605</guid>
		</item>
		<item>
			<title><![CDATA[Re: Offline version of the Magenta Audio to Midi transcription tool]]></title>
			<link>https://forum.modartt.com/viewtopic.php?pid=979602#p979602</link>
			<description><![CDATA[<div class="quotebox"><cite>Vagporto wrote:</cite><blockquote><p>I very much doubt that a model for automatic pedal detection can ever be created.</p><p>I totally agree with you...</p></blockquote></div><p>respectfully i must disagree.&nbsp; or at least i have to say, these arguments are not convincing to me.</p><p>the point is the way these models work.&nbsp; they are doing absolutely nothing like the kind of quantitative reasoning a human might do to evaluate whether the pedal has been depressed.&nbsp; all they do is try to design an algorithm by analyzing lots and lots of data that has already been marked as &quot;pedal depressed&quot; or &quot;pedal not depressed&quot;.&nbsp; the software makes a guess, scores itself, and then improves its guess.&nbsp; the mathematics behind the process is very interesting and can&#039;t really be gone through here, but the point is there is a well-defined way for the software to modify its algorithm based on its performance.&nbsp; after many many iterations of this, assuming the model is designed well, it will perform very well on the training set, and also on new input it&#039;s never seen before.&nbsp; how it&#039;s &quot;thinking&quot; is completely different from how a human actually thinks.</p><p>already a lot of the sonic ambiguity mentioned is present in solo piano and is handled very well by onsets and frames.&nbsp; for instance, how can we distinguish the fundamental of a note and some higher overtone?&nbsp; the tone of a piano note is very different when it&#039;s played loudly or softly.&nbsp; how do we distinguish that?&nbsp; we don&#039;t ... we just train the model on data and let it sort it out internally.&nbsp; (to be fair, one weakness of the current model is detecting velocity.&nbsp; the midi velocity values are reasonable but not close to a human performance, imo.&nbsp; still it&#039;s doing a remarkable job.)</p><p>of course the only way to demonstrate success is succeed.&nbsp; i&#039;d like to try at some point, but i don&#039;t know when i&#039;ll have time.&nbsp; maybe someone else will try <i class="far fa-smile smiley"></i></p>]]></description>
			<author><![CDATA[null@example.com (budo)]]></author>
			<pubDate>Sun, 19 Dec 2021 22:32:54 +0000</pubDate>
			<guid>https://forum.modartt.com/viewtopic.php?pid=979602#p979602</guid>
		</item>
		<item>
			<title><![CDATA[Re: Offline version of the Magenta Audio to Midi transcription tool]]></title>
			<link>https://forum.modartt.com/viewtopic.php?pid=979588#p979588</link>
			<description><![CDATA[<div class="quotebox"><cite>Vagporto wrote:</cite><blockquote><div class="quotebox"><cite>teacue wrote:</cite><blockquote><p>@ YvesTh<br />Very interesting points, thank you for this.<br />A not easy task and probably even more difficult for a non-pianist.</p><p>@ budo<br />A new magenta model would be indeed welcome.</p></blockquote></div><p>I very much doubt that a model for automatic pedal detection can ever be created.<br />From a sonic perspective, it is impossible to differentiate if a note is sustained by the key or by the pedal (yes, the resonances are different if other notes are played at the same time, but I bet my money that the machine learning model focuses only on fundamental frequencies or otherwise the analysis would be incredibly complex). Moreover, you have the sustain pedal that lifts all dampers, and the sostenuto pedal that prevents lifted dampers from dropping. So, when a note is held, it is impossible to know which of the three methods is being used:&nbsp; key, sustain pedal, sostenuto pedal.</p><p>The only way is applying a &quot;feasibility&quot; test: each finger is doing what? Does this requires more than 10 fingers?</p></blockquote></div><p>I totally agree with you...</p>]]></description>
			<author><![CDATA[null@example.com (YvesTh)]]></author>
			<pubDate>Sun, 19 Dec 2021 14:46:36 +0000</pubDate>
			<guid>https://forum.modartt.com/viewtopic.php?pid=979588#p979588</guid>
		</item>
		<item>
			<title><![CDATA[Re: Offline version of the Magenta Audio to Midi transcription tool]]></title>
			<link>https://forum.modartt.com/viewtopic.php?pid=979587#p979587</link>
			<description><![CDATA[<div class="quotebox"><cite>teacue wrote:</cite><blockquote><p>@ YvesTh<br />Very interesting points, thank you for this.<br />A not easy task and probably even more difficult for a non-pianist.</p><p>@ budo<br />A new magenta model would be indeed welcome.</p></blockquote></div><p>I very much doubt that a model for automatic pedal detection can ever be created.<br />From a sonic perspective, it is impossible to differentiate if a note is sustained by the key or by the pedal (yes, the resonances are different if other notes are played at the same time, but I bet my money that the machine learning model focuses only on fundamental frequencies or otherwise the analysis would be incredibly complex). Moreover, you have the sustain pedal that lifts all dampers, and the sostenuto pedal that prevents lifted dampers from dropping. So, when a note is held, it is impossible to know which of the three methods is being used:&nbsp; key, sustain pedal, sostenuto pedal.</p><p>The only way is applying a &quot;feasibility&quot; test: each finger is doing what? Does this requires more than 10 fingers?</p>]]></description>
			<author><![CDATA[null@example.com (Vagporto)]]></author>
			<pubDate>Sun, 19 Dec 2021 14:13:44 +0000</pubDate>
			<guid>https://forum.modartt.com/viewtopic.php?pid=979587#p979587</guid>
		</item>
		<item>
			<title><![CDATA[Re: Offline version of the Magenta Audio to Midi transcription tool]]></title>
			<link>https://forum.modartt.com/viewtopic.php?pid=979439#p979439</link>
			<description><![CDATA[<p>@ YvesTh<br />Thank you for your thoughts</p>]]></description>
			<author><![CDATA[null@example.com (teacue)]]></author>
			<pubDate>Tue, 14 Dec 2021 14:32:47 +0000</pubDate>
			<guid>https://forum.modartt.com/viewtopic.php?pid=979439#p979439</guid>
		</item>
		<item>
			<title><![CDATA[Re: Offline version of the Magenta Audio to Midi transcription tool]]></title>
			<link>https://forum.modartt.com/viewtopic.php?pid=979386#p979386</link>
			<description><![CDATA[<p>Midi file extract :</p><p>With pedal sustain addition at the begining and various corrections :</p><p><a href="https://forum.modartt.com/uploads.php?file=Duke%20a%20goutelas%20%28extrait%20fichier%20midi%29.mid">https://forum.modartt.com/uploads.php?f...idi%29.mid</a><br />With velocity curve :<br />Global Velocity = [0, 26, 46, 72, 102; 0, 23, 44, 79, 127]<br />Duke Ellington was playing on a Steinway for this recording.<br />Pictures on this link :<br /><a href="https://ellington.se/2021/02/25/ellington-at-chateau-goutelas-1966/">https://ellington.se/2021/02/25/ellingt...elas-1966/</a></p>]]></description>
			<author><![CDATA[null@example.com (YvesTh)]]></author>
			<pubDate>Sun, 12 Dec 2021 16:25:57 +0000</pubDate>
			<guid>https://forum.modartt.com/viewtopic.php?pid=979386#p979386</guid>
		</item>
		<item>
			<title><![CDATA[Re: Offline version of the Magenta Audio to Midi transcription tool]]></title>
			<link>https://forum.modartt.com/viewtopic.php?pid=979384#p979384</link>
			<description><![CDATA[<div class="quotebox"><cite>teacue wrote:</cite><blockquote><p>You both did not answer my question about dynamic.<br />Any idea about this?</p></blockquote></div><p>You are right, on my test the dynamic needs to be reworked. I think that the choice of the pianoteq instrument used is very important, to that it is necessary to rework the velocity curve for the reading of the midi file and even sometimes to modify individually the velocity of some notes.</p>]]></description>
			<author><![CDATA[null@example.com (YvesTh)]]></author>
			<pubDate>Sun, 12 Dec 2021 16:01:30 +0000</pubDate>
			<guid>https://forum.modartt.com/viewtopic.php?pid=979384#p979384</guid>
		</item>
		<item>
			<title><![CDATA[Re: Offline version of the Magenta Audio to Midi transcription tool]]></title>
			<link>https://forum.modartt.com/viewtopic.php?pid=979383#p979383</link>
			<description><![CDATA[<p>@ YvesTh<br />Very interesting points, thank you for this.<br />A not easy task and probably even more difficult for a non-pianist.</p><p>@ budo<br />A new magenta model would be indeed welcome.</p><p>You both did not answer my question about dynamic.<br />Any idea about this?</p>]]></description>
			<author><![CDATA[null@example.com (teacue)]]></author>
			<pubDate>Sun, 12 Dec 2021 15:31:53 +0000</pubDate>
			<guid>https://forum.modartt.com/viewtopic.php?pid=979383#p979383</guid>
		</item>
		<item>
			<title><![CDATA[Re: Offline version of the Magenta Audio to Midi transcription tool]]></title>
			<link>https://forum.modartt.com/viewtopic.php?pid=979351#p979351</link>
			<description><![CDATA[<div class="quotebox"><cite>teacue wrote:</cite><blockquote><p>As a non-pianist I would like to learn in this context how to restore the sustain pedal.<br />Has someone some tips?</p></blockquote></div><p>@YvesTh gave a great answer with everything i wanted to say about this.&nbsp; i just wanted to add that a longer-term way to solve the problem would be to develop a new magenta model that does this.&nbsp; it would take as input the same data as onsets and frames (wav file) and produce a guess about application of the sustain pedal.&nbsp; it could be trained on lots of audio snippets, the same as onsets and frames.&nbsp; it should actually be easier to do this than what onsets and frames already does, because it&#039;s training for a binary state (pedal is on vs pedal is off).&nbsp; of course it could be made more complicated because there are lots of coloristic effects one can do with the pedal that go beyond on/off, but at least this would be a start.</p>]]></description>
			<author><![CDATA[null@example.com (budo)]]></author>
			<pubDate>Sat, 11 Dec 2021 19:46:20 +0000</pubDate>
			<guid>https://forum.modartt.com/viewtopic.php?pid=979351#p979351</guid>
		</item>
		<item>
			<title><![CDATA[Re: Offline version of the Magenta Audio to Midi transcription tool]]></title>
			<link>https://forum.modartt.com/viewtopic.php?pid=979347#p979347</link>
			<description><![CDATA[<div class="quotebox"><cite>teacue wrote:</cite><blockquote><p>As a non-pianist I would like to learn in this context how to restore the sustain pedal.<br />Has someone some tips?</p></blockquote></div><p>If the notes held are unplayable by the pianist (finger spread) it means that the sustain pedal has been used. Otherwise you have to listen carefully to the original file to see if you can hear the particular resonance of the piano with the pedal depressed. The problem is that pianists often use the pedal on short sequences of notes that are very difficult to determine, and the time and speed of releasing the pedal is just as important as the time and speed of depressing it. I have on the &quot;Duke&quot; file a lot of trouble to determine everywhere these parameters....<br />For classical music you can read the score indications but for a jazz improvisation you have only your ears...</p><p>Translated with <a href="http://www.DeepL.com/Translator">www.DeepL.com/Translator</a> (free version)</p>]]></description>
			<author><![CDATA[null@example.com (YvesTh)]]></author>
			<pubDate>Sat, 11 Dec 2021 13:25:04 +0000</pubDate>
			<guid>https://forum.modartt.com/viewtopic.php?pid=979347#p979347</guid>
		</item>
		<item>
			<title><![CDATA[Re: Offline version of the Magenta Audio to Midi transcription tool]]></title>
			<link>https://forum.modartt.com/viewtopic.php?pid=979267#p979267</link>
			<description><![CDATA[<p>@ YvesTh<br />A nice and interesting transcription of Satin Doll <i class="far fa-smile smiley"></i></p><p>As a non-pianist I would like to learn in this context how to restore the sustain pedal.<br />Has someone some tips?</p><p>As one can clearly hear in YvesTh transcription much of the dynamic get lost.<br />Beside of editing each note manually what could be done to get the whole dynamic range?<br />I tried editing the velocity curve but found out that it is not easy to achieve good results.</p>]]></description>
			<author><![CDATA[null@example.com (teacue)]]></author>
			<pubDate>Wed, 08 Dec 2021 16:03:09 +0000</pubDate>
			<guid>https://forum.modartt.com/viewtopic.php?pid=979267#p979267</guid>
		</item>
		<item>
			<title><![CDATA[Re: Offline version of the Magenta Audio to Midi transcription tool]]></title>
			<link>https://forum.modartt.com/viewtopic.php?pid=979258#p979258</link>
			<description><![CDATA[<p>Thanks, Yves.&nbsp; I have gone from &#039;wowed by the technology&#039; to now having glimpsed behind the curtain and seen all the work that must go on behind the scenes.&nbsp; Clearly, it&#039;s not &#039;turn-key&#039; technology the way that shooting a digital photo of an old print or slide is, where auto-exposure makes that pretty easy now-a-days.&nbsp; I am sure that an &#039;auto-musicality&#039; function can&#039;t be too far behind!&nbsp; ;-)</p>]]></description>
			<author><![CDATA[null@example.com (dklein)]]></author>
			<pubDate>Wed, 08 Dec 2021 02:49:26 +0000</pubDate>
			<guid>https://forum.modartt.com/viewtopic.php?pid=979258#p979258</guid>
		</item>
		<item>
			<title><![CDATA[Re: Offline version of the Magenta Audio to Midi transcription tool]]></title>
			<link>https://forum.modartt.com/viewtopic.php?pid=979252#p979252</link>
			<description><![CDATA[<p>Here is my test of magenta :</p><p>Original file : Duke Ellington solo piano at Goutelas in 1966</p><p><a href="https://forum.modartt.com/uploads.php?file=Duke%20at%20Goutelas%201966%20%28extrait%29.mp3">https://forum.modartt.com/uploads.php?f...ait%29.mp3</a></p><p>Pianoteq File &quot;Steinway B&quot; with midi file from magenta ( with some corrections).</p><p><a href="https://forum.modartt.com/uploads.php?file=Duke%20at%20Goutelas%202021%28in%20progress%29.mp3">https://forum.modartt.com/uploads.php?f...ess%29.mp3</a></p><p>I will try to make many other necessary corrections.</p><p>Main problems on the midi file.<br />Many excess notes had to be removed. <br /> Low notes coming from Duke&#039;s grunts, high notes coming from very strong harmonics, high notes coming from clapping, the software not being able to detect the sustain pedal, the notes are extended individually and therefore we don&#039;t have the sympathetic resonance of the other strings with pianoteq. The ideal would be to restore the pedal effects in the midi file (I added some quickly). Some missing notes too. It&#039;s will be a big work to do but very interesting.<br />The original file is about 2 x 10 minutes long.</p>]]></description>
			<author><![CDATA[null@example.com (YvesTh)]]></author>
			<pubDate>Tue, 07 Dec 2021 21:21:14 +0000</pubDate>
			<guid>https://forum.modartt.com/viewtopic.php?pid=979252#p979252</guid>
		</item>
	</channel>
</rss>
