<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>http://groupkos.com/dev/index.php?action=history&amp;feed=atom&amp;title=Principles_of_Sinusoidal_Positioning</id>
	<title>Principles of Sinusoidal Positioning - Revision history</title>
	<link rel="self" type="application/atom+xml" href="http://groupkos.com/dev/index.php?action=history&amp;feed=atom&amp;title=Principles_of_Sinusoidal_Positioning"/>
	<link rel="alternate" type="text/html" href="http://groupkos.com/dev/index.php?title=Principles_of_Sinusoidal_Positioning&amp;action=history"/>
	<updated>2026-05-15T16:50:52Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.39.3</generator>
	<entry>
		<id>http://groupkos.com/dev/index.php?title=Principles_of_Sinusoidal_Positioning&amp;diff=3614&amp;oldid=prev</id>
		<title>XenoEngineer: Created page with &quot;{{menuLMPrimer}} &lt;div style=&quot;background-color:azure; border:1px outset azure; padding:0 20px; max-width:860px; margin:0 auto; &quot;&gt; ===Principles of Sinusoidal Positioning=== ====Periodicity, Frequency, Orthogonality====  In natural language processing (NLP) and deep learning, positional encoding is a technique used to incorporate information about the position of each token in a sequence into a vector representation. This allows the model to capture sequential relationship...&quot;</title>
		<link rel="alternate" type="text/html" href="http://groupkos.com/dev/index.php?title=Principles_of_Sinusoidal_Positioning&amp;diff=3614&amp;oldid=prev"/>
		<updated>2024-08-09T11:50:42Z</updated>

		<summary type="html">&lt;p&gt;Created page with &amp;quot;{{menuLMPrimer}} &amp;lt;div style=&amp;quot;background-color:azure; border:1px outset azure; padding:0 20px; max-width:860px; margin:0 auto; &amp;quot;&amp;gt; ===Principles of Sinusoidal Positioning=== ====Periodicity, Frequency, Orthogonality====  In natural language processing (NLP) and deep learning, positional encoding is a technique used to incorporate information about the position of each token in a sequence into a vector representation. This allows the model to capture sequential relationship...&amp;quot;&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;{{menuLMPrimer}}&lt;br /&gt;
&amp;lt;div style=&amp;quot;background-color:azure; border:1px outset azure; padding:0 20px; max-width:860px; margin:0 auto; &amp;quot;&amp;gt;&lt;br /&gt;
===Principles of Sinusoidal Positioning===&lt;br /&gt;
====Periodicity, Frequency, Orthogonality====&lt;br /&gt;
&lt;br /&gt;
In natural language processing (NLP) and deep learning, positional encoding is a technique used to incorporate information about the position of each token in a sequence into a vector representation. This allows the model to capture sequential relationships between tokens, even when they don&amp;#039;t have direct connections.&lt;br /&gt;
&lt;br /&gt;
One popular method for positional encoding is sinusoidal positional encoding, introduced by Vaswani et al. in their 2017 paper &amp;quot;Attention is All You Need&amp;quot;. The idea is to use sine and cosine functions to create a vector representation that captures the position of each token.&lt;br /&gt;
&lt;br /&gt;
The SPE formula is as follows:&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;code&amp;gt;&lt;br /&gt;
positional_encoding = sin(10000^(2i/sequence_length)) + cos(10000^(2i/sequence_length))&lt;br /&gt;
&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;big&amp;gt;&amp;#039;&amp;#039;&amp;#039;i&amp;#039;&amp;#039;&amp;#039;&amp;lt;/big&amp;gt; is the index of the token in the sequence&amp;lt;br/&amp;gt;&lt;br /&gt;
 &amp;lt;big&amp;gt;&amp;#039;&amp;#039;&amp;#039;sequence_length&amp;#039;&amp;#039;&amp;#039;&amp;lt;/big&amp;gt; is the length of the input sequence&lt;br /&gt;
&lt;br /&gt;
The resulting vector has a fixed dimension, typically set to 512. The sine and&lt;br /&gt;
cosine functions are evaluated at different frequencies for each token&lt;br /&gt;
position, creating a unique vector representation.&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;1. Periodicity:&amp;#039;&amp;#039;&amp;#039; SPE uses periodic functions (sine and cosine) to capture the position of tokens in a sequence.&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;2. Frequency:&amp;#039;&amp;#039;&amp;#039; The frequency of the sine and cosine functions increases as you move along the sequence, allowing the model to distinguish between different positions.&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;2. Orthogonality:&amp;#039;&amp;#039;&amp;#039; The vectors generated by SPE are orthogonal to each other, which helps prevent overfitting and improves generalization.&lt;br /&gt;
&lt;br /&gt;
By incorporating positional information into a vector representation, SPE enables models like transformers to learn complex relationships between tokens in a sequence.&lt;br /&gt;
&amp;lt;/div&amp;gt;&lt;/div&gt;</summary>
		<author><name>XenoEngineer</name></author>
	</entry>
</feed>