<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Alesandro Slepčević - Architect the cloud</title>
	<atom:link href="https://blog.slepcevic.net/author/admin/feed/" rel="self" type="application/rss+xml" />
	<link>https://blog.slepcevic.net</link>
	<description></description>
	<lastBuildDate>Mon, 06 Oct 2025 09:23:48 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9</generator>
	<item>
		<title>Your own AI coding assistant running on Akamai cloud!</title>
		<link>https://blog.slepcevic.net/your-own-ai-coding-assistant-running-on-akamai-cloud/</link>
					<comments>https://blog.slepcevic.net/your-own-ai-coding-assistant-running-on-akamai-cloud/#respond</comments>
		
		<dc:creator><![CDATA[Alesandro Slepčević]]></dc:creator>
		<pubDate>Tue, 18 Feb 2025 18:17:12 +0000</pubDate>
				<category><![CDATA[Akamai Connected Cloud]]></category>
		<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[Cloud]]></category>
		<category><![CDATA[Open Source]]></category>
		<category><![CDATA[Terraform]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[AI coding]]></category>
		<category><![CDATA[bolt.diy]]></category>
		<category><![CDATA[Linode]]></category>
		<category><![CDATA[Linux]]></category>
		<guid isPermaLink="false">https://blog.slepcevic.net/?p=760</guid>

					<description><![CDATA[<p>What? You want some AI to write my code? AI-powered coding assistants are the main talk in the developer world for a while, there&#8217;s no denying that. I can&#8217;t count the times I&#8217;ve read somewhere the AI will replace developers...</p>
<p>The post <a href="https://blog.slepcevic.net/your-own-ai-coding-assistant-running-on-akamai-cloud/">Your own AI coding assistant running on Akamai cloud!</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></description>
										<content:encoded><![CDATA[<h3 class="wp-block-heading">What? You want some AI to write my code?</h3>



<p>AI-powered coding assistants are the main talk in the developer world for a while, there&#8217;s no denying that. I can&#8217;t count the times I&#8217;ve read somewhere the AI will replace developers in the next X years. You’ve probably seen tools like <strong>GitHub Copilot, ChatGPT, or Tabnine</strong> popping up everywhere.</p>



<p>They promise to boost productivity, help with debugging, and even teach you new coding techniques. Sounds amazing, right? But like anything, AI-powered coding assistants have their downsides too. So, let’s talk about what makes them great—and where they might fall short.</p>



<h3 class="wp-block-heading">Why AI Coding Assistants Are a Game-Changer</h3>



<p>Obviously, one of the biggest advantages of using an AI assistant is the time it saves. Instead of writing the same repetitive boilerplate code over and over and over again, you can generate it in seconds. Need a quick function to parse JSON? AI has you covered.  Easy peasy! Stuck on how to structure your SQL query? Just ask. This means less time spent on the boring stuff and more time on actual problem-solving.</p>



<p>AI is also a fantastic debugging tool. It can analyze your code, catch potential issues, and suggest fixes before you even run it. Instead of spending hours combing through error messages and Stack Overflow threads, you get quick, relevant suggestions that help you move forward faster.</p>



<p>And let’s not forget about learning. If you’re picking up a new language or framework, an AI assistant can guide you with real-time examples, explain unfamiliar syntax, and even generate sample projects. It’s like having a 24/7 coding mentor who doesn’t judge your questions.</p>



<p>Beyond just speed and learning, AI can actually help improve code quality. It can suggest best practices, helps format your code, and even recommends refactoring when your code gets messy. Plus, if you’re working in a team, it can assist with keeping code style consistent and even generate useful commit messages or documentation. Wouldn&#8217;t it be cool if we could plug the AI into our pipeline and make sure that all rules are being followed?</p>



<h3 class="wp-block-heading">The downsides no one(everyone) talks about?</h3>



<p>As cool as AI coding assistants are, you don&#8217;t need to be a genius to see that they&#8217;re far from perfect. One of the biggest concerns I personally see is over-reliance. If you’re constantly relying on AI to write your code, do you really understand what’s happening under the hood? This can be a problem when something breaks, and you don’t know how to fix it because you never really wrote the thing in the first place <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f642.png" alt="🙂" class="wp-smiley" style="height: 1em; max-height: 1em;" /> I&#8217;m sure you love reading someone else&#8217;s codebase and debugging that &lt;3</p>



<p>Another issue is that AI-generated code isn’t always optimized or even correct! It might suggest something that works but isn’t efficient, secure, or maintainable. If you blindly accept AI suggestions without reviewing them, you could end up with a mess of inefficient or buggy code.</p>



<p>Then there’s the question of security. AI assistants are trained on huge datasets, and sometimes they can generate code that includes security vulnerabilities. If you’re working on sensitive stuff, you have to be extra careful about what code you’re using and where it’s coming from.</p>



<p>Let&#8217;s talk privacy! Many AI coding tools rely on cloud-based processing, meaning your code might be sent to external servers for analysis. If you’re working on proprietary or confidential code, you need to be aware of the risks and check the privacy policies of the tools you’re using.</p>



<p>And finally, while AI can make you more productive, it can also be a bit of a crutch. Some developers might start relying too much on AI for even basic things, which can slow down their growth and problem-solving skills in the long run. </p>



<h3 class="wp-block-heading">So, Should You Use One?</h3>



<p>AI coding assistants are undeniably powerful tools, but they work best when used wisely. They’re great for boosting productivity, helping with debugging, and learning new technologies—but they shouldn’t replace actual coding knowledge and problem-solving skills. Think of them as a really smart assistant, not a replacement for your own expertise.</p>



<p>If you use AI responsibly—review its suggestions, stay mindful of security risks, and make sure you’re still learning and improving as a developer it can be a fantastic addition to your workflow, just don’t let it do all the thinking for you <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f642.png" alt="🙂" class="wp-smiley" style="height: 1em; max-height: 1em;" /></p>



<h3 class="wp-block-heading">Still interested and want to start using AI in your daily work? Enter bolt.diy <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f642.png" alt="🙂" class="wp-smiley" style="height: 1em; max-height: 1em;" /></h3>



<p><strong>bolt.diy</strong> is the open source version of Bolt.new (previously known as oTToDev and bolt.new ANY LLM), which allows you to choose the LLM that you use for each prompt! Currently, you can use OpenAI, Anthropic, Ollama, OpenRouter, Gemini, LMStudio, Mistral, xAI, HuggingFace, DeepSeek, or Groq models&nbsp;</p>



<figure class="wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio"><div class="wp-block-embed__wrapper">
<div class="responsive-embed widescreen"><iframe title="I Forked Bolt.new and Made it WAY Better" width="1000" height="563" src="https://www.youtube.com/embed/3PFcAu_oU80?list=PLyrg3m7Ei-MpOPKdenkQNcx8ueI36RNrA" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" referrerpolicy="strict-origin-when-cross-origin" allowfullscreen></iframe></div>
</div></figure>



<p><strong>bolt.diy</strong> was originally started by&nbsp;<a href="https://www.youtube.com/@ColeMedin">Cole Medin</a>&nbsp;but has quickly grown into a massive community effort to build the one of the open source AI coding assistants out there.</p>



<h2 class="wp-block-heading">What do I need to get this deployed?</h2>



<p>Well, just Terraform and a Linode account. <br>In the backend we will deploy a VM with a GPU attached, install bolt.diy, ollama and ask it to write some code! Maybe a simple Tic-Tac-Toe game? <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f642.png" alt="🙂" class="wp-smiley" style="height: 1em; max-height: 1em;" /></p>



<p>Ideally you would run your bolt.diy deployment on a separate machine from the machine running the model, but for our use case, current deployment model is more than enough.</p>



<p>Like most of the things on this blog, guess what we&#8217;re gonna use? Yes! IaC!!! <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f600.png" alt="😀" class="wp-smiley" style="height: 1em; max-height: 1em;" /></p>



<p>Here&#8217;s a link to the <a href="https://github.com/aslepcev/linode-bolt.diy" target="_blank" rel="noopener" title="">Github repository</a> containing the Terraform code. </p>



<p>Code will do the following:</p>



<ol class="wp-block-list">
<li>Deploy a GPU based instance in Akamai Connected Cloud</li>



<li>Use cloud-init to install the following:
<ul class="wp-block-list">
<li>curl</li>



<li>wget</li>



<li>nodejs</li>



<li>npm</li>



<li>nvtop &#8211; great tool to monitor your GPU usage</li>



<li>Nvidia drivers</li>
</ul>
</li>



<li>Deploy and configure a firewall which will allow SSH and <strong>bolt.diy</strong> access from your IP. </li>



<li>Configure <strong>bolt </strong>and <strong>ollama </strong>to run as a Linux service. For <strong>ollama </strong>service, we are always making sure we have a model downloaded and created with 32K context size. </li>
</ol>



<h2 class="wp-block-heading">How do you deploy it? </h2>



<p>Just fill in your Linode API token and the desired region, Linode token and your IP address in <strong>variables.tf </strong>file and run the following commands:</p>



<pre class="wp-block-code"><code>git clone https://github.com/aslepcev/linode-bolt.diy
cd linode-bolt.diy
#Fill in the variables.tf file now
terrafom init
terraform plan
terraform apply</code></pre>



<p>After a short 5-6 minute wait, everything should be deployed and ready to use. Go ahead and visit the IP address of your VM on the port 5173. </p>



<p><strong>Example url</strong>: http://172.233.246.209:5173</p>



<p>Make sure that Ollama is selected as a provider and you&#8217;re off to the races!</p>



<h2 class="wp-block-heading">What can it do? </h2>



<p>Well, it really depends on the model we are running. With the <strong><a href="https://www.linode.com/pricing/#compute-gpu" target="_blank" rel="noopener" title="">RTX 4000 Ada GPU</a></strong>, we can comfortably run a <strong>14B parameter model with 32K context size</strong> which is &#8220;ok&#8221; for smaller and simpler stuff. </p>



<p>I tested it out with a simple task of creating a Tic-Tac-Toe game in NodeJS. It got the functionality right the first time, but it looked like something only a mother could love <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f642.png" alt="🙂" class="wp-smiley" style="height: 1em; max-height: 1em;" /></p>



<figure class="wp-block-image size-large"><img fetchpriority="high" decoding="async" width="1024" height="505" src="https://blog.slepcevic.net/wp-content/uploads/2025/02/bolt1-1024x505.png" alt="" class="wp-image-770" srcset="https://blog.slepcevic.net/wp-content/uploads/2025/02/bolt1-1024x505.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2025/02/bolt1-300x148.png 300w, https://blog.slepcevic.net/wp-content/uploads/2025/02/bolt1-768x379.png 768w, https://blog.slepcevic.net/wp-content/uploads/2025/02/bolt1-1536x758.png 1536w, https://blog.slepcevic.net/wp-content/uploads/2025/02/bolt1.png 1615w" sizes="(max-width: 1024px) 100vw, 1024px" /></figure>



<p>I just told it to make it a bit prettier and add some color; these were the results I got:</p>



<figure class="wp-block-image size-full"><img decoding="async" width="694" height="589" src="https://blog.slepcevic.net/wp-content/uploads/2025/02/bolt-3.png" alt="" class="wp-image-771" srcset="https://blog.slepcevic.net/wp-content/uploads/2025/02/bolt-3.png 694w, https://blog.slepcevic.net/wp-content/uploads/2025/02/bolt-3-300x255.png 300w" sizes="(max-width: 694px) 100vw, 694px" /></figure>



<p>Interestingly, during the coding process, it made a mistake which it managed to identify and fix all on its own! All I did was press the &#8220;<strong>Ask Bolt</strong>&#8221; button. </p>



<p></p>



<figure class="wp-block-image size-full is-resized"><img loading="lazy" decoding="async" width="616" height="251" src="https://blog.slepcevic.net/wp-content/uploads/2025/02/bolt-fix.png" alt="" class="wp-image-772" style="width:838px;height:auto" srcset="https://blog.slepcevic.net/wp-content/uploads/2025/02/bolt-fix.png 616w, https://blog.slepcevic.net/wp-content/uploads/2025/02/bolt-fix-300x122.png 300w" sizes="auto, (max-width: 616px) 100vw, 616px" /></figure>



<p>Also, here&#8217;s a fully functioning Space Invaders alike game which it also wrote</p>



<figure class="wp-block-image size-full"><img loading="lazy" decoding="async" width="975" height="665" src="https://blog.slepcevic.net/wp-content/uploads/2025/02/bolt-space.png" alt="" class="wp-image-774" srcset="https://blog.slepcevic.net/wp-content/uploads/2025/02/bolt-space.png 975w, https://blog.slepcevic.net/wp-content/uploads/2025/02/bolt-space-300x205.png 300w, https://blog.slepcevic.net/wp-content/uploads/2025/02/bolt-space-768x524.png 768w" sizes="auto, (max-width: 975px) 100vw, 975px" /></figure>



<h2 class="wp-block-heading">What if I want to run a larger model? 32B parameters or even larger? </h2>



<p>That&#8217;s very easy! Since Ollama can use multiple GPU&#8217;s, all we need to do is scale up the VM we are using to the one which includes two or more GPU&#8217;s. Akamai offers maximum of 4 GPU&#8217;s per VM which brings up to 80 GB of VRAM which we can use to run our model. I will not experiment with larger models in this blog post; this is something we will benchmark and try out in the future. </p>



<p>Cheers! Alex.</p>



<p>P.S &#8211; parts of this post were written by bolt.diy <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f642.png" alt="🙂" class="wp-smiley" style="height: 1em; max-height: 1em;" /></p>



<p></p><p>The post <a href="https://blog.slepcevic.net/your-own-ai-coding-assistant-running-on-akamai-cloud/">Your own AI coding assistant running on Akamai cloud!</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></content:encoded>
					
					<wfw:commentRss>https://blog.slepcevic.net/your-own-ai-coding-assistant-running-on-akamai-cloud/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>DeepSeek-R1 on Akamai Connected Cloud</title>
		<link>https://blog.slepcevic.net/deepseek-r1-on-akamai-connected-cloud-linode/</link>
					<comments>https://blog.slepcevic.net/deepseek-r1-on-akamai-connected-cloud-linode/#comments</comments>
		
		<dc:creator><![CDATA[Alesandro Slepčević]]></dc:creator>
		<pubDate>Tue, 28 Jan 2025 10:01:00 +0000</pubDate>
				<category><![CDATA[Akamai Connected Cloud]]></category>
		<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[Linux]]></category>
		<category><![CDATA[Open Source]]></category>
		<guid isPermaLink="false">https://blog.slepcevic.net/?p=718</guid>

					<description><![CDATA[<p>Before we dive into installation and real usage, let&#8217;s talk about DeepSeek-R1 a bit and why it&#8217;s a big thing. The team behind DeepSeek-R1-Zero took a cool approach to advancing large language models by skipping supervised fine-tuning (SFT) entirely and...</p>
<p>The post <a href="https://blog.slepcevic.net/deepseek-r1-on-akamai-connected-cloud-linode/">DeepSeek-R1 on Akamai Connected Cloud</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></description>
										<content:encoded><![CDATA[<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="823" height="402" src="https://blog.slepcevic.net/wp-content/uploads/2025/01/deepseeklogo-edited.png" alt="" class="wp-image-742" srcset="https://blog.slepcevic.net/wp-content/uploads/2025/01/deepseeklogo-edited.png 823w, https://blog.slepcevic.net/wp-content/uploads/2025/01/deepseeklogo-edited-300x147.png 300w, https://blog.slepcevic.net/wp-content/uploads/2025/01/deepseeklogo-edited-768x375.png 768w" sizes="auto, (max-width: 823px) 100vw, 823px" /></figure>



<p><strong>Before we dive into installation and real usage, let&#8217;s talk about DeepSeek-R1 a bit and why it&#8217;s a big thing</strong>. </p>



<p>The team behind DeepSeek-R1-Zero took a cool approach to advancing large language models by skipping supervised fine-tuning (SFT) entirely and jumping straight into reinforcement learning (RL). </p>



<p>This resulted in a model capable of self-verification, reflection, and generating intricate chain-of-thoughts (CoTs) to solve complex problems. What’s borderline groundbreaking here is that this is the first open research to show RL alone can incentivize reasoning in LLMs, without the need for SFT. Development pipeline for DeepSeek-R1 includes two RL stages to improve reasoning patterns and align with human preferences, along with two SFT stages to seed the model&#8217;s reasoning and non-reasoning capabilities. </p>



<p>They didn’t stop at advancing reasoning in large models—they also focused on making smaller models more powerful! By distilling the reasoning patterns of larger models into smaller ones, they achieved better performance than what RL alone could achieve on small models. </p>



<p>Using DeepSeek-R1’s reasoning data, they fine-tuned several dense models and demonstrated exceptional benchmark results. To further benefit the community, they’ve open-sourced distilled checkpoints ranging from 1.5B to 70B parameters, based on the Qwen2.5 and Llama3 series.</p>



<p><strong>What is distillation you may ask?</strong></p>



<p>In the context of AI, <strong>distillation</strong> (short for <em>knowledge distillation</em>) is a technique used to transfer knowledge from a larger, more complex model (called the <em>teacher</em>) to a smaller, simpler model (called the <em>student</em>). The goal is to make the smaller model perform as well as—or almost as well as—the larger model, while requiring significantly fewer resources (e.g., memory, computation).</p>



<h3 class="wp-block-heading">How Distillation Works:</h3>



<ol class="wp-block-list">
<li><strong>Training the Teacher Model:</strong><br>A large model is trained first. This model is usually highly expressive, capable of capturing complex patterns in the data, but it may be too computationally intensive for practical deployment.</li>



<li><strong>Soft Labels as Knowledge:</strong><br>Instead of just training the student model on the original dataset, the student learns from the <em>soft predictions</em> of the teacher model. These predictions include probability distributions over all possible outcomes (rather than binary correct/incorrect labels), which contain richer information about the relationships between classes.</li>



<li><strong>Student Model Training:</strong><br>The student model is trained to mimic the teacher’s behavior using a combination of:
<ul class="wp-block-list">
<li>The soft predictions from the teacher model (knowledge transfer).</li>



<li>The original ground-truth data.</li>
</ul>
</li>
</ol>



<h3 class="wp-block-heading">Why Distillation is Useful:</h3>



<ul class="wp-block-list">
<li><strong>Efficiency:</strong> Smaller models are faster, require less memory, and can run on devices with limited computational power (like mobile phones or edge devices).</li>



<li><strong>Scalability:</strong> By distilling large models into smaller ones, researchers and developers can deploy AI more broadly while maintaining strong performance.</li>



<li><strong>Accessibility:</strong> Distilled models often serve as the foundation for open-source projects, making advanced AI capabilities more accessible to the community.</li>
</ul>



<p>Ok, let&#8217;s get to the interesting part and deploy our model. We can break down the entire deployment in 4 easy steps.</p>



<p><strong>Step 1</strong> &#8211; Deploy a GPU VM in Linode. I&#8217;m gonna deploy the smallest GPU VM available &#8220;RTX4000 Ada x1 Small&#8221; which costs 350$ per month. </p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="554" src="https://blog.slepcevic.net/wp-content/uploads/2025/01/Screenshot-2025-01-29-100730-1024x554.png" alt="" class="wp-image-719" srcset="https://blog.slepcevic.net/wp-content/uploads/2025/01/Screenshot-2025-01-29-100730-1024x554.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2025/01/Screenshot-2025-01-29-100730-300x162.png 300w, https://blog.slepcevic.net/wp-content/uploads/2025/01/Screenshot-2025-01-29-100730-768x415.png 768w, https://blog.slepcevic.net/wp-content/uploads/2025/01/Screenshot-2025-01-29-100730.png 1413w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<p><strong>Step 2</strong> &#8211; Connect using SSH to your instance and install Nvidia drivers</p>



<p>Since I&#8217;m using Debian 12 as my OS, installing Nvidia drivers was quite easy following official instructions &#8211; <a href="https://wiki.debian.org/NvidiaGraphicsDrivers" target="_blank" rel="noopener" title="">https://wiki.debian.org/NvidiaGraphicsDrivers</a></p>



<p>After the driver has been install, I suggest you also install &#8220;nvtop&#8221; using: </p>



<pre class="wp-block-code"><code>apt install nvtop</code></pre>



<p></p>



<p><strong>nvtop </strong>is a cool utility which can monitor the usage of your GPU. Think of it like a &#8220;top&#8221; command, but for GPU&#8217;s. </p>



<p><strong>Step 3</strong> &#8211; Install Ollama</p>



<p>Installing Ollama is extremely simple, all you need to do it run a single command. </p>



<pre class="wp-block-code"><code>curl -fsSL https://ollama.com/install.sh | sh</code></pre>



<p></p>



<p><strong>Step 4</strong> &#8211; Download and run the DeepSeek-R1 model into Ollama. Since the GPU I will be using has 20 GB of VRAM, we can fit a 14B model without any issues. </p>



<p>&#8211;verbose flag tells Ollama to display performance metrics after the generation has been completed. </p>



<pre class="wp-block-code"><code>ollama run deepseek-r1:14b --verbose
</code></pre>



<p></p>



<p>Let&#8217;s ask a simple question and see how it performs: <strong>What is distilling in the context of AI?</strong></p>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow">
<p><strong>&lt;think&gt;</strong></p>



<p>Hmm, distillation… I think it&#8217;s related to machine learning, specifically in training<br>models. Maybe it&#8217;s about transferring knowledge from one model to another? Let me recall.<br>Oh right, model distillation is the process where you take a complex, large model and<br>create a smaller, more efficient version that retains the same performance.</p>



<p>Wait, there are different types of distillation too. I should mention those. There&#8217;s<br>knowledge distillation where you transfer knowledge from a teacher model to a student<br>model. Then there&#8217;s parameter-efficient methods like fine-tuning or adapter-based<br>approaches. Also, pruning and quantization techniques come under distilling in terms of<br>optimizing models.</p>



<p>I need to make sure the explanation is clear but not too technical. The user might be<br>someone new to AI or looking to understand optimization techniques. They probably want a<br>practical understanding without getting bogged down by jargon.</p>



<p>Maybe I should start with the general idea, then break it down into types, and explain<br>each briefly. Also, highlighting why it&#8217;s important—like efficiency, scalability,<br>deployment on devices—would be helpful for them to grasp its value.</p>



<p>I should keep the response structured but conversational, avoiding any markdown. Just<br>plain, easy-to-understand English. Let me put that together.<br><strong>&lt;/think&gt;</strong></p>



<p>In the context of AI and machine learning, <strong>distilling</strong> refers to a technique where<br>knowledge or information from one model is transferred to another, often simpler or<br>smaller model. This process aims to create a more efficient version of the original model<br>while preserving its performance.</p>



<h3 class="wp-block-heading">Key Ideas:</h3>



<ol class="wp-block-list">
<li><strong>Model Distillation</strong>:</li>
</ol>



<ul class="wp-block-list">
<li>It involves training a &#8220;student&#8221; model to mimic the behavior or decisions of a<br>larger, more complex &#8220;teacher&#8221; model.</li>



<li>The teacher model is typically pre-trained on a large dataset and has high accuracy<br>but may be computationally expensive or difficult to deploy.</li>
</ul>



<ol class="wp-block-list">
<li><strong>Types of Distillation</strong>:</li>
</ol>



<ul class="wp-block-list">
<li><strong>Knowledge Distillation</strong>: The teacher model&#8217;s predictions (soft labels) are used to<br>train the student model, which can lead to better generalization compared to using only<br>hard labels.</li>



<li><strong>Parameter-Efficient Methods</strong>: Techniques like fine-tuning, adapter-based<br>approaches, or prompt tuning allow the transfer of knowledge without copying all<br>parameters from the teacher model.</li>



<li><strong>Pruning and Quantization</strong>: These are techniques used to reduce the size of a model<br>by removing unnecessary weights (pruning) or reducing the precision of the weights<br>(quantization).</li>
</ul>



<ol class="wp-block-list">
<li><strong>Applications</strong>:</li>
</ol>



<ul class="wp-block-list">
<li>Distillation is widely used to deploy AI models on edge devices where computational<br>resources are limited.</li>



<li>It also helps in scaling down large models for faster inference while maintaining<br>performance.</li>
</ul>



<p>In summary, distilling in AI is about making complex models more efficient and accessible<br>by transferring their knowledge to simpler or smaller models.</p>



<p></p>
</blockquote>



<pre class="wp-block-code"><code>total duration: 19.004002094s
load duration: 12.267119ms
prompt eval count: 105 token(s)
prompt eval duration: 48ms
prompt eval rate: 2187.50 tokens/s
eval count: 615 token(s)
eval duration: 18.938s
<strong>eval rate: 32.47 tokens/s</strong></code></pre>



<p></p>



<p>One major difference between DeepSeek-R1 and majority of other generative AI models is that DeepSeek will show you the &#8220;train of thought&#8221; it used to get to the answer, not just the answer itself. </p>



<p></p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="364" src="https://blog.slepcevic.net/wp-content/uploads/2025/01/ollama3-1024x364.png" alt="" class="wp-image-720" srcset="https://blog.slepcevic.net/wp-content/uploads/2025/01/ollama3-1024x364.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2025/01/ollama3-300x107.png 300w, https://blog.slepcevic.net/wp-content/uploads/2025/01/ollama3-768x273.png 768w, https://blog.slepcevic.net/wp-content/uploads/2025/01/ollama3-1536x546.png 1536w, https://blog.slepcevic.net/wp-content/uploads/2025/01/ollama3.png 1570w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<p>Once we have the model running locally, we can turn it into our own personal AI assistant by using <a href="https://github.com/open-webui/open-webui" target="_blank" rel="noopener" title="Open-WebUI">Open-WebUI</a> or turn it into your own coding assistant using <a href="https://github.com/stackblitz-labs/bolt.diy" target="_blank" rel="noopener" title="Bolt.diy">Bolt.diy</a>. </p>



<p>Ollama also has it&#8217;s on <a href="https://github.com/ollama/ollama/blob/main/docs/api.md" target="_blank" rel="noopener" title="API">API</a> which you can use to embed the model into your own applications.</p>



<p><strong>Possibilities are endless!</strong></p>



<p>BTW, we can even run a <strong>32B parameter model,</strong> but the performance drops to 10 tokens per second and spills a bit into the RAM and CPU, but nothing drastic. Model is still fully usable and comparable to ChatGPT on a busy day.  </p>



<p>Results you can see on the image are based on the request to build me &#8220;Space Invaders&#8221; game in HTML. </p>



<figure class="wp-block-image size-full is-resized"><img loading="lazy" decoding="async" width="792" height="394" src="https://blog.slepcevic.net/wp-content/uploads/2025/01/deepseek32b.png" alt="" class="wp-image-734" style="width:840px;height:auto" srcset="https://blog.slepcevic.net/wp-content/uploads/2025/01/deepseek32b.png 792w, https://blog.slepcevic.net/wp-content/uploads/2025/01/deepseek32b-300x149.png 300w, https://blog.slepcevic.net/wp-content/uploads/2025/01/deepseek32b-768x382.png 768w" sizes="auto, (max-width: 792px) 100vw, 792px" /></figure>



<p>Until next time where we will take a look into bolt.diy and deploy our own coding assistant!</p>



<p>Alex!</p>



<p></p><p>The post <a href="https://blog.slepcevic.net/deepseek-r1-on-akamai-connected-cloud-linode/">DeepSeek-R1 on Akamai Connected Cloud</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></content:encoded>
					
					<wfw:commentRss>https://blog.slepcevic.net/deepseek-r1-on-akamai-connected-cloud-linode/feed/</wfw:commentRss>
			<slash:comments>2</slash:comments>
		
		
			</item>
		<item>
		<title>Akamai Application Platform on LKE</title>
		<link>https://blog.slepcevic.net/akamai-application-platform-on-lke/</link>
					<comments>https://blog.slepcevic.net/akamai-application-platform-on-lke/#respond</comments>
		
		<dc:creator><![CDATA[Alesandro Slepčević]]></dc:creator>
		<pubDate>Thu, 24 Oct 2024 00:06:40 +0000</pubDate>
				<category><![CDATA[Akamai APL]]></category>
		<category><![CDATA[Akamai Application Platform]]></category>
		<category><![CDATA[Akamai Connected Cloud]]></category>
		<category><![CDATA[APL]]></category>
		<category><![CDATA[Otomi]]></category>
		<category><![CDATA[Terraform]]></category>
		<guid isPermaLink="false">https://blog.slepcevic.net/?p=674</guid>

					<description><![CDATA[<p>The Akamai Application Platform is a Kubernetes-centric solution that supports the full application lifecycle from development to deployment. It leverages tools from the Cloud Native Computing Foundation (CNCF) landscape, providing an integrated environment for managing containerized workloads. Optimized for Linode...</p>
<p>The post <a href="https://blog.slepcevic.net/akamai-application-platform-on-lke/">Akamai Application Platform on LKE</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></description>
										<content:encoded><![CDATA[<p>The <strong>Akamai Application Platform</strong> is a Kubernetes-centric solution that supports the full application lifecycle from development to deployment. It leverages tools from the Cloud Native Computing Foundation (CNCF) landscape, providing an integrated environment for managing containerized workloads. </p>



<p>Optimized for Linode Kubernetes Engine (LKE), the platform enhances automation and self-service, streamlining application management without requiring custom-built Kubernetes solutions.</p>



<p>Let&#8217;s get it up and running; like everything I do, we will avoid clickops and use IaC. </p>



<p>I&#8217;ve put the entire codebase on Github so you can simply clone it, configure your Linode token and run Terraform apply.</p>



<p>Code will deploy a LKE cluster in Amsterdam with 3 x <strong>Linode 8 GB</strong> instances and then use HELM provider to deploy Akamai Application Platform for you. </p>



<p>It will also create a DNS zone with the domain specified in the code and configure APL with full DNS integration, meaning you&#8217;ll be pretty much ready to go and deploy apps. </p>



<p><strong>Please note that the cluster might scale up in case you decide to deploy a lot of applications. </strong></p>



<p><strong>Step 1 </strong>&#8211; clone the repository</p>



<pre class="wp-block-code"><code>git clone https://github.com/slepix/ApplicationPlatform-Linode.git</code></pre>



<p></p>



<p><strong>Step 2</strong> &#8211; fill out apl.tfvars file</p>



<pre class="wp-block-code"><code>linode_token = "YourLinodeToken"
domain = "yourdomain.com"
region = "nl-ams"
k8s_version = "1.30"
soaemail = "SoaEmail@yourdomain.com"</code></pre>



<p></p>



<p><strong>Step 3</strong> &#8211; run terraform plan to verify what will be deployed</p>



<pre class="wp-block-code"><code>terraform plan --var-file="apl.tfvars"</code></pre>



<p></p>



<p><strong>Step 4 </strong>&#8211; run Terraform apply and grab some coffee</p>



<pre class="wp-block-code"><code>terraform apply --var-file="apl.tfvars"</code></pre>



<p></p>



<p>After the cluster has been deployed; you can go to Linode console and open the Kubernetes dashboard (you will need to download the kubeconfig file as well). </p>



<p>If all went smooth (and it will, because IaC :D), you should see a bunch of containers popping up while APL configures itself. </p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="411" src="https://blog.slepcevic.net/wp-content/uploads/2024/10/apl1-1024x411.png" alt="" class="wp-image-679" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/10/apl1-1024x411.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/10/apl1-300x120.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/10/apl1-768x308.png 768w, https://blog.slepcevic.net/wp-content/uploads/2024/10/apl1-1536x616.png 1536w, https://blog.slepcevic.net/wp-content/uploads/2024/10/apl1.png 1968w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<p>Click on Jobs on the left hand side and then on the three dots next to &#8220;apl&#8221; job. </p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="426" src="https://blog.slepcevic.net/wp-content/uploads/2024/10/apl3-1024x426.png" alt="" class="wp-image-680" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/10/apl3-1024x426.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/10/apl3-300x125.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/10/apl3-768x320.png 768w, https://blog.slepcevic.net/wp-content/uploads/2024/10/apl3-1536x639.png 1536w, https://blog.slepcevic.net/wp-content/uploads/2024/10/apl3.png 1656w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<p>Click on Logs, which will bring up a new window with the logs.  Once the installation has completed, you will see the username, password and the URL where you can log in after everything has been set up. </p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="641" src="https://blog.slepcevic.net/wp-content/uploads/2024/10/apl2-1024x641.png" alt="" class="wp-image-681" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/10/apl2-1024x641.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/10/apl2-300x188.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/10/apl2-768x480.png 768w, https://blog.slepcevic.net/wp-content/uploads/2024/10/apl2.png 1159w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<p>Have in mind that setup will take a good 20-30 mins; there&#8217;s a lot of things which needs to be configured in the background for you <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f642.png" alt="🙂" class="wp-smiley" style="height: 1em; max-height: 1em;" /></p>



<p>After everything has been deployed, the console website should become accessible. Login with the credentials listed in the logs of your job and you should be presented with the console. </p>



<figure class="wp-block-image size-full"><img loading="lazy" decoding="async" width="990" height="694" src="https://blog.slepcevic.net/wp-content/uploads/2024/10/apl5.png" alt="" class="wp-image-682" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/10/apl5.png 990w, https://blog.slepcevic.net/wp-content/uploads/2024/10/apl5-300x210.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/10/apl5-768x538.png 768w" sizes="auto, (max-width: 990px) 100vw, 990px" /></figure>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="519" src="https://blog.slepcevic.net/wp-content/uploads/2024/10/apl6-1024x519.png" alt="" class="wp-image-684" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/10/apl6-1024x519.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/10/apl6-300x152.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/10/apl6-768x389.png 768w, https://blog.slepcevic.net/wp-content/uploads/2024/10/apl6-1536x779.png 1536w, https://blog.slepcevic.net/wp-content/uploads/2024/10/apl6.png 1619w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<p><strong>Make sure to follow post-installation steps mentioned <a href="https://apl-docs.net/docs/get-started/installation/post-installation-steps" target="_blank" rel="noopener" title="">here</a> in order to finish the configuration.</strong></p>



<p>Cheers, </p>



<p>Alex!</p><p>The post <a href="https://blog.slepcevic.net/akamai-application-platform-on-lke/">Akamai Application Platform on LKE</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></content:encoded>
					
					<wfw:commentRss>https://blog.slepcevic.net/akamai-application-platform-on-lke/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Get your own own Github runner deployed and configured on Linode in less than 5 minutes.</title>
		<link>https://blog.slepcevic.net/get-your-own-own-github-runner-deployed-and-configured-on-linode-in-less-than-5-minutes/</link>
					<comments>https://blog.slepcevic.net/get-your-own-own-github-runner-deployed-and-configured-on-linode-in-less-than-5-minutes/#respond</comments>
		
		<dc:creator><![CDATA[Alesandro Slepčević]]></dc:creator>
		<pubDate>Wed, 23 Oct 2024 23:19:20 +0000</pubDate>
				<category><![CDATA[Akamai Connected Cloud]]></category>
		<category><![CDATA[Architecture]]></category>
		<category><![CDATA[Cloud]]></category>
		<category><![CDATA[GitHub]]></category>
		<category><![CDATA[GitHub runner]]></category>
		<category><![CDATA[Linux]]></category>
		<guid isPermaLink="false">https://blog.slepcevic.net/?p=685</guid>

					<description><![CDATA[<p>Why would you run your own runner anyways? GitHub Actions (along with Azure DevOps) has emerged as a powerful managed tool that allows developers to automate workflows directly within their GitHub repositories. While GitHub provides hosted runners to execute these...</p>
<p>The post <a href="https://blog.slepcevic.net/get-your-own-own-github-runner-deployed-and-configured-on-linode-in-less-than-5-minutes/">Get your own own Github runner deployed and configured on Linode in less than 5 minutes.</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></description>
										<content:encoded><![CDATA[<p><strong>Why would you run your own runner anyways?</strong></p>



<p>GitHub Actions (along with Azure DevOps) has emerged as a powerful managed tool that allows developers to automate workflows directly within their GitHub repositories. While GitHub provides hosted runners to execute these workflows, running your own GitHub runner can offer several advantages. </p>



<p><strong>1. Money talks</strong></p>



<p>One of the primary benefits of running your own GitHub runner is cost efficiency. GitHub Actions provides a certain number of free minutes for public and private repositories, but once you exceed these limits, costs can add up quickly, especially if your execution takes a while <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f641.png" alt="🙁" class="wp-smiley" style="height: 1em; max-height: 1em;" /></p>



<p><strong>2</strong>. <strong>Customization and Control</strong></p>



<p>When you run your own GitHub runner, you gain full control over the environment in which your workflows execute. This means you can customize the runner&#8217;s operating system, software, and dependencies to match your project’s specific needs. It doesn&#8217;t matter anymore if you need a particular version of a programming language, specialized libraries, or specific system configurations, your self-hosted runner can be tailored to your requirements.</p>



<p><strong>3</strong>. <strong>Performance and Speed</strong></p>



<p>Self-hosted runners can significantly enhance the performance of your CI/CD pipelines. Since these runners are dedicated to your projects, you can optimize them for speed and efficiency. You can run builds on beefy machines, use faster storage, or even set up parallel execution across multiple runners to speed up your workflows. This is especially beneficial for larger projects or teams with multiple repositories or a bunch of members working in parallel. </p>



<p><strong>4. Security and Compliance</strong></p>



<p>For organizations handling sensitive data or operating in regulated industries, security is number one. Running your own GitHub runner allows you to maintain control over your CI/CD environment. You can implement your own security measures, restrict network access, and ensure that sensitive information does not leave your secured infrastructure. Additionally, you can regularly update and audit your runner to comply with internal policies or external regulations.</p>



<p><strong>5. Reduced Queue Times</strong></p>



<p>Using GitHub&#8217;s hosted runners means you may encounter queue times, especially during peak usage periods. By setting up your own runners, you can mitigate these delays, ensuring that your workflows kick off as soon as possible. </p>



<p><strong>How do I get it running? </strong></p>



<p><strong>Step 1</strong> &#8211; Clone the repository using the following command. </p>



<pre class="wp-block-code"><code>git clone https://github.com/slepix/GitHubRunner-Linode.git</code></pre>



<p></p>



<h5 class="wp-block-heading">You will need to prepare 5 things; it&#8217;s not hard, I promise. </h5>



<ol class="wp-block-list">
<li><strong>Linode API token</strong> with permission to deploy virtual machines &#8211; <a href="https://techdocs.akamai.com/cloud-computing/docs/manage-personal-access-tokens" target="_blank" rel="noopener" title="">more info</a></li>



<li><strong>GitHub PAT</strong> limited only to the repository you want to connect &#8211;<a href="https://docs.github.com/en/authentication/keeping-your-account-and-data-secure/managing-your-personal-access-tokens" target="_blank" rel="noopener" title=""> more details</a></li>



<li><strong>GitHub repository</strong> name you want to connect your runner to. </li>



<li><strong>GitHub username</strong> which owns the repository. </li>



<li><strong>Root password</strong> for the runner VM &#8211; can be anything, as long as it&#8217;s long and complex. </li>
</ol>



<p><strong>Step 2</strong> &#8211; Fill in the details in <strong>linoderunner.tfvars</strong> file; it should look something like this. </p>



<p>*These are random values, so make sure to replace them with your own. </p>



<pre class="wp-block-code"><code>linode_api_token = "eee44387b0030bd6bb051452bg65gz56z465fba5d77c5a238ea8e12f"
github_pat = "github_pat_1HG534e67d3K52IUSL_D2vM1pzjDGjX5sCiUEXWD6TRKDut4jnJty"
root_password = "Rand0mSecurePassword.123!" # Root password for your VM
github_repo = "myawesomeapp" # Your Github repo name
github_username = "slepix" # Your GitHub username</code></pre>



<p><strong>Step 3</strong> &#8211; run the following command:</p>



<pre class="wp-block-code"><code>terraform apply --var-file="linoderunner.tfvars"</code></pre>



<p>Entire codebase is available at <a href="https://github.com/slepix/GithubRunner-Linode" target="_blank" rel="noopener" title="">https://github.com/slepix/GithubRunner-Linode</a></p>



<p>Ok, let&#8217;s take a look at some code. Once again, we&#8217;ll go with Terraform and cloud-init to deploy and configure our server. Ideally you would use some configuration management tool like Puppet, Ansible, Chef or similar, but for this use case, we can keep it simple. </p>



<p>Using cloud-init, we create a new user called &#8220;<strong>gitrunner</strong>&#8221; which will be used to run the agent, update all the packes, install jq (needed by the agent configuration script) and kick off the installation of the runner as a service. </p>



<p><strong>compute.tf file</strong> &#8211; this is where you can adjust the region, OS and instance type you want to run. </p>



<pre class="wp-block-code"><code>resource "linode_instance" "github_runner" {
  image     = "linode/ubuntu22.04"
  region    = "nl-ams"
  type      = "g6-nanode-1"
  label     = "github-runner"
  root_pass = var.root_password  # Set the root password

  metadata {
    user_data = base64encode(templatefile("./linode.yaml.tpl", {
      githubpat = var.github_pat
      githubuser = var.github_username
      githubrepo = var.github_repo
    }))
  }
}</code></pre>



<p><strong>linode.yaml.tpl file</strong></p>



<pre class="wp-block-code"><code>#cloud-config
package_update: true
packages:
  - jq

users:
  - name: gitrunner
    shell: /bin/bash
    groups:
      - sudo
    sudo:
      - ALL=(ALL) NOPASSWD:ALL

runcmd:
  - export RUNNER_CFG_PAT=${githubpat}
  - su gitrunner -c "cd /home/gitrunner &amp;&amp; curl -s https://raw.githubusercontent.com/actions/runner/main/scripts/create-latest-svc.sh | bash -s ${githubuser}/${githubrepo}" 
</code></pre>



<p>If all went good, you should see the new GitHub runner appear in your runner overview in a few mins. </p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="316" src="https://blog.slepcevic.net/wp-content/uploads/2024/10/runner-1024x316.png" alt="" class="wp-image-686" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/10/runner-1024x316.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/10/runner-300x92.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/10/runner-768x237.png 768w, https://blog.slepcevic.net/wp-content/uploads/2024/10/runner.png 1152w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<p>Happy building and deploying!</p>



<p>Alex!</p><p>The post <a href="https://blog.slepcevic.net/get-your-own-own-github-runner-deployed-and-configured-on-linode-in-less-than-5-minutes/">Get your own own Github runner deployed and configured on Linode in less than 5 minutes.</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></content:encoded>
					
					<wfw:commentRss>https://blog.slepcevic.net/get-your-own-own-github-runner-deployed-and-configured-on-linode-in-less-than-5-minutes/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Linode, Windows instances and Metadata service</title>
		<link>https://blog.slepcevic.net/linode-windows-instances-and-metadata-service/</link>
					<comments>https://blog.slepcevic.net/linode-windows-instances-and-metadata-service/#respond</comments>
		
		<dc:creator><![CDATA[Alesandro Slepčević]]></dc:creator>
		<pubDate>Tue, 30 Jul 2024 13:22:17 +0000</pubDate>
				<category><![CDATA[Akamai Connected Cloud]]></category>
		<category><![CDATA[Cloud]]></category>
		<category><![CDATA[Powershell]]></category>
		<category><![CDATA[Windows]]></category>
		<category><![CDATA[Linode]]></category>
		<category><![CDATA[windows linode metadata]]></category>
		<guid isPermaLink="false">https://blog.slepcevic.net/?p=662</guid>

					<description><![CDATA[<p>Even though Windows instances aren&#8217;t officially supported on Linode, there are ways to get a Windows based instance up and running (a bit of googling will help you around that). Once you deployed your Windows instance, you will notice that...</p>
<p>The post <a href="https://blog.slepcevic.net/linode-windows-instances-and-metadata-service/">Linode, Windows instances and Metadata service</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></description>
										<content:encoded><![CDATA[<p>Even though Windows instances aren&#8217;t officially supported on Linode, there are ways to get a Windows based instance up and running (a bit of googling will help you around that). </p>



<p>Once you deployed your Windows instance, you will notice that all attempts to reach the Linode&#8217;s <a href="https://www.linode.com/docs/products/compute/compute-instances/guides/metadata/?tabs=linode-api" target="_blank" rel="noopener" title="metadata service">metadata service</a> will fail. </p>



<p>Let&#8217;s take a quick look on how metadata service works. </p>



<pre class="wp-block-code"><code>#We get the token into a TOKEN variable
export TOKEN=$(curl -X PUT -H "Metadata-Token-Expiry-Seconds: 3600" http://169.254.169.254/v1/token)

#Send a request to the metadata service. 
curl -H "Metadata-Token: $TOKEN" http://169.254.169.254/v1/instance</code></pre>



<p></p>



<p>If we translate these commands into PowerShell, we get something which looks like this</p>



<pre class="wp-block-code"><code># Define the URL for the token request
$tokenUrl = "http://169.254.169.254/v1/token"

# Define the headers for the token request
$tokenHeaders = @{
    "Metadata-Token-Expiry-Seconds" = "3600"
}

# Make the PUT request to get the token
$tokenResponse = Invoke-RestMethod -Uri $tokenUrl -Method Put -Headers</code></pre>



<p></p>



<p>If we execute Invoke-RestMethod command out of the box, the request will fail with the following message: Unable to connect to the remote server</p>



<figure class="wp-block-image size-full"><img loading="lazy" decoding="async" width="848" height="145" src="https://blog.slepcevic.net/wp-content/uploads/2024/07/windowsmetadata.png" alt="" class="wp-image-663" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/07/windowsmetadata.png 848w, https://blog.slepcevic.net/wp-content/uploads/2024/07/windowsmetadata-300x51.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/07/windowsmetadata-768x131.png 768w" sizes="auto, (max-width: 848px) 100vw, 848px" /></figure>



<p>Problem occurs because our server doesn&#8217;t know where to route the request for http://169.254.169.254 and can easily be fixed by adding a route on your server which will route requests for 169.254.169.254 via your instance gateway. </p>



<pre class="wp-block-code"><code>ROUTE ADD 169.254.169.254 mask 255.255.255.255 <strong>172.233.45.1</strong> METRIC 10</code></pre>



<p></p>



<p>Make sure to replace <strong>172.233.45.1</strong> with your gateway IP address. Don&#8217;t use the one provided above. After adding the route, we should be able to send requests to our metadata service. Below you can find an example PowerShell script which will return basic instance information. </p>



<pre class="wp-block-code"><code># Define the URL for the token request
$tokenUrl = "http://169.254.169.254/v1/token"

# Define the headers for the token request
$tokenHeaders = @{
    "Metadata-Token-Expiry-Seconds" = "3600"
}

# Make the PUT request to get the token
$tokenResponse = Invoke-RestMethod -Uri $tokenUrl -Method Put -Headers $tokenHeaders 

# Extract the token from the response
$token = $tokenResponse

# Define the URL for the instance info request
$networkUrl = "http://169.254.169.254/v1/instance"

# Define the headers for the network request
$networkHeaders = @{
    "Metadata-Token" = $token
    "Accept" = "application/json" #specify we want the results in JSON format
}

# Make the GET request to retrieve instance information
$Response = Invoke-RestMethod -Uri $networkUrl -Method Get -Headers $networkHeaders 

return $Response
</code></pre>



<p></p>



<p>Finally, if you want the route for the metadata service to be persistent across reboots, make sure to add the &#8220;-p&#8221; flag to your &#8220;route add&#8221; command. </p>



<p>Cheers, </p>



<p>Alex. </p><p>The post <a href="https://blog.slepcevic.net/linode-windows-instances-and-metadata-service/">Linode, Windows instances and Metadata service</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></content:encoded>
					
					<wfw:commentRss>https://blog.slepcevic.net/linode-windows-instances-and-metadata-service/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Powershell module for Linode &#8211; Akamai Connected Cloud</title>
		<link>https://blog.slepcevic.net/powershell-module-for-linode-akamai-connected-cloud/</link>
					<comments>https://blog.slepcevic.net/powershell-module-for-linode-akamai-connected-cloud/#respond</comments>
		
		<dc:creator><![CDATA[Alesandro Slepčević]]></dc:creator>
		<pubDate>Sun, 30 Jun 2024 13:09:53 +0000</pubDate>
				<category><![CDATA[Akamai Connected Cloud]]></category>
		<category><![CDATA[Cloud]]></category>
		<category><![CDATA[Linux]]></category>
		<category><![CDATA[Open Source]]></category>
		<category><![CDATA[Powershell]]></category>
		<category><![CDATA[Windows]]></category>
		<category><![CDATA[Linode]]></category>
		<category><![CDATA[Linode PowerShell Module]]></category>
		<category><![CDATA[Powershell module]]></category>
		<guid isPermaLink="false">https://blog.slepcevic.net/?p=487</guid>

					<description><![CDATA[<p>Even though I consider myself as being &#8220;platform-agnostic&#8221;, I originally come from a Microsoft side of the industry and because of that I had a lot of exposure to PowerShell. After a relatively steep learning curve, I realized that PowerShell...</p>
<p>The post <a href="https://blog.slepcevic.net/powershell-module-for-linode-akamai-connected-cloud/">Powershell module for Linode – Akamai Connected Cloud</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></description>
										<content:encoded><![CDATA[<p>Even though I consider myself as being &#8220;platform-agnostic&#8221;, I originally come from a Microsoft side of the industry and because of that I had a lot of exposure to PowerShell. </p>



<p>After a relatively steep learning curve, I realized that PowerShell is cool and VERY powerful for some use cases. </p>



<p></p>



<p>Well, as a small side project, I started working on a <a href="https://github.com/slepix/LinodePowerShell" target="_blank" rel="noopener" title="">Powershell module</a> for Akamai Connected Cloud (Linode) platform.</p>



<p><strong>Why? Why not!? </strong></p>



<p>Linode is already known as being <a href="https://www.linode.com/developers/" target="_blank" rel="noopener" title="">VERY developer friendly</a>, why not make it even more friendlier <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f600.png" alt="😀" class="wp-smiley" style="height: 1em; max-height: 1em;" /></p>



<p>General idea was to generate a module which will directly talk to the Linode API and eventually expose all platform features via some commandlet. </p>



<p>Besides that, I also wanted to implement some quality of life features which current Linode CLI doesn&#8217;t have.</p>



<h5 class="wp-block-heading">Using the label/name of the resource in your commands.</h5>



<p>With the current <a href="https://www.linode.com/docs/products/tools/cli/get-started/" target="_blank" rel="noopener" title="">linode-cli</a>, you need to use the resource ID in all commands which is a bit annoying for some use cases <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f642.png" alt="🙂" class="wp-smiley" style="height: 1em; max-height: 1em;" /></p>



<p>With this module, you can use the resource label/name in your commands, and in the background the module will convert the label provided to a resource ID and use that to talk to the API. </p>



<h5 class="wp-block-heading">Support for multiple Linode accounts and account switching.</h5>



<p>I also implemented the possibility to have and switch between multiple different Linode &#8220;profiles&#8221;. </p>



<p>You can have a profile for your development Linode account, a profile for your test Linode account and switch between them using a single command. </p>



<h5 class="wp-block-heading">Generate a random password when launching an instance</h5>



<p>When you deploy instances with the linode-cli, you need to provide a password for the instance. </p>



<p>Since humans are lazy, we most probably end up using the same password over and over again. This feature will make sure that you always get a secure password generated for your Linode. Of course, you are still free to provide your own password if you like. </p>



<h6 class="wp-block-heading"><strong>Check out the examples below to see how all of these things work. </strong></h6>



<p></p>



<h4 class="wp-block-heading"><strong>How to start? It&#8217;s really easy! You will be up and running in less than 2 minutes!</strong></h4>



<h4 class="wp-block-heading">Installation</h4>



<p>To install the Linode PowerShell module, you need to clone the repository, position yourself into the module folder and import the module:</p>



<pre class="wp-block-code"><code>git clone https://github.com/slepix/LinodePowerShell.git
cd LinodePowerShell</code></pre>



<p></p>



<h4 class="wp-block-heading">Usage</h4>



<p>Before you can use the module, you need to import it using the following command:</p>



<pre class="wp-block-code"><code>Import-Module -Name .\LinodePSModule.psd1</code></pre>



<p>After the module has been imported, you need to configure it and add your Linode account.</p>



<h4 class="wp-block-heading">Configuration</h4>



<p>Configuration is really simple, all you need to do is run&nbsp;<strong>&#8216;Connect-LinodeAccount&#8217;</strong>&nbsp;command and follow the configuration wizard.</p>



<pre class="wp-block-code"><code>Connect-LinodeAccount</code></pre>



<p></p>



<p><strong>You can run </strong>&#8220;Connect-LinodeAccount&#8221; <strong>command for each Linode account you want to add.</strong></p>



<pre class="wp-block-preformatted">IMPORTANT!!!<br>When creating a new profile, you will be asked if you want to encrypt your Linode token or not. If you decide to encrypt the token, you won't be able to manually edit the profile file and edit the token. <br><br>It is recommended that you use encrypted profile option and recreate a profile if you need to change the token.<br></pre>



<p>User configuration is stored in &#8220;<strong>$HOME\.LinodePSModule\$name-profile.json</strong>&#8220;</p>



<p>Check in the examples below how you can switch between different Linode accounts.</p>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow">
<p><em><strong>NOTE:</strong></em><br><strong>You will need Linode API token with the READ/WRITE permissions for all resources in order to get the full functionality of this module.<br>However, you are free to limit the scope of the API token to suit your security and usage requirements</strong>.</p>
</blockquote>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="350" src="https://blog.slepcevic.net/wp-content/uploads/2024/06/blogpsmodule-1024x350.png" alt="" class="wp-image-609" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/06/blogpsmodule-1024x350.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/06/blogpsmodule-300x103.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/06/blogpsmodule-768x262.png 768w, https://blog.slepcevic.net/wp-content/uploads/2024/06/blogpsmodule.png 1191w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<p></p>



<h4 class="wp-block-heading">Available commands</h4>



<p><a href="https://github.com/slepix/LinodePowerShell#commands"></a></p>



<p>List of currently available commands can be viewed by running this command</p>



<pre class="wp-block-code"><code>(Get-Module LinodePSModule).ExportedCommands</code></pre>



<p></p>



<h4 class="wp-block-heading">Example commands</h4>



<p><a href="https://github.com/slepix/LinodePowerShell#example-commands"></a></p>



<p><strong>List all currently configured profiles</strong></p>



<pre class="wp-block-code"><code>PS C:\&gt; Get-LinodeProviderProfiles # returns a list of all configured profiles
development
test</code></pre>



<p></p>



<p><strong>Switch to a profile named &#8220;test&#8221;</strong></p>



<pre class="wp-block-code"><code>PS C:\&gt; Set-LinodeProviderProfile -profile test
Profile test loaded</code></pre>



<p></p>



<p><strong>Get a list of all regions, but only return label, id and status of each region</strong></p>



<pre class="wp-block-code"><code>PS C:\&gt; Get-LinodeRegions | Select label, id, status

label           id           status
-----           --           ------
Mumbai, IN      ap-west      ok
Toronto, CA     ca-central   ok
Sydney, AU      ap-southeast ok
Washington, DC  us-iad       ok
Chicago, IL     us-ord       ok
Paris, FR       fr-par       ok
Seattle, WA     us-sea       ok
Sao Paulo, BR   br-gru       ok
Amsterdam, NL   nl-ams       ok
Stockholm, SE   se-sto       ok
Madrid, ES      es-mad       ok
Chennai, IN     in-maa       ok
Osaka, JP       jp-osa       ok
Milan, IT       it-mil       ok
Miami, FL       us-mia       ok
Jakarta, ID     id-cgk       ok
Los Angeles, CA us-lax       ok
Dallas, TX      us-central   ok
Fremont, CA     us-west      ok
Atlanta, GA     us-southeast ok
Newark, NJ      us-east      ok
London, UK      eu-west      ok
Singapore, SG   ap-south     ok
Frankfurt, DE   eu-central   ok
Tokyo, JP       ap-northeast ok</code></pre>



<p></p>



<p></p>



<p><strong>Returns all properties of a single region</strong></p>



<pre class="wp-block-code"><code>PS C:\&gt; Get-LinodeRegion -region nl-ams # list only nl-ams region

id                     : nl-ams
label                  : Amsterdam, NL
country                : nl
capabilities           : {Linodes, Backups, NodeBalancers, Block Storage...}
status                 : ok
resolvers              : @{ipv4=172.233.33.36, 172.233.33.38, 172.233.33.35,     172.233.33.39, 172.233.33.34, 172.233.33.33, 172.233.33.31, 172.233.33.30, 172.233.33.37, 172.233.33.32;
                         ipv6=2600:3c0e::f03c:93ff:fe9d:2d10, 2600:3c0e::f03c:93ff:fe9d:2d89, 2600:3c0e::f03c:93ff:fe9d:2d79, 2600:3c0e::f03c:93ff:fe9d:2d96, 2600:3c0e::f03c:93ff:fe9d:2da5,     
                         2600:3c0e::f03c:93ff:fe9d:2d34, 2600:3c0e::f03c:93ff:fe9d:2d68, 2600:3c0e::f03c:93ff:fe9d:2d17, 2600:3c0e::f03c:93ff:fe9d:2d45, 2600:3c0e::f03c:93ff:fe9d:2d5c}
placement_group_limits : @{maximum_pgs_per_customer=100; maximum_linodes_per_pg=5}
site_type              : core</code></pre>



<p></p>



<p>Create a new Linode named &#8220;<strong>myInstance</strong>&#8221; with the instance type of &#8220;<strong>g6-nanode-1</strong>&#8221; in <strong>nl-ams</strong> region and a <strong>randomly generated password</strong> <strong>25 characters long</strong> running <strong>Debian 11</strong></p>



<pre class="wp-block-code"><code>PS C:\&gt; New-LinodeInstance -label myInstance -region nl-ams -type g6-nanode-1 -image linode/debian11 -generatepassword -passwordlength 25


id               : 60827598
label            : myInstance
group            :
status           : provisioning
created          : 2024-06-30T01:23:46
updated          : 2024-06-30T01:23:46
type             : g6-nanode-1
ipv4             : {172.233.40.46}
ipv6             : 2600:3c0e::f03c:94ff:fe85:29cf/128
image            : linode/debian11
region           : nl-ams
specs            : @{disk=25600; memory=1024; vcpus=1; gpus=0; transfer=1000}
alerts           : @{cpu=90; network_in=10; network_out=10; transfer_quota=80; io=10000}
backups          : @{enabled=False; available=False; schedule=; last_successful=}
hypervisor       : kvm
watchdog_enabled : True
tags             : {}
host_uuid        : 5f337dda11805445fc391c91e6dcbf45c2a38a21
has_user_data    : False
placement_group  :
lke_cluster_id   :
root_password    : eJbd\OQC9ypSagfvTSG@XREV4</code></pre>



<p></p>



<p><strong>Delete a Linode instance</strong></p>



<p><strong>All destructive commands</strong> will force you to use &#8220;-confirm&#8221; flag in order to actually delete a resource.</p>



<pre class="wp-block-code"><code>PS C:\&gt; Remove-LinodeInstance -label myInstance  
Please use the -confirm flag in order to delete an instance</code></pre>



<p></p>



<pre class="wp-block-code"><code>PS C:\&gt; Remove-LinodeInstance -label myInstance -confirm
Instance 60827598 deleted</code></pre>



<p></p>



<h5 class="wp-block-heading">Every other command</h5>



<p>Even though each commandlet will ask you for parameters, you can also use&nbsp;<em><strong>&#8216;Get-Help&#8217;</strong></em>&nbsp;commandlet in order to find out which parameters each command requires.</p>



<p><strong>Example for &#8220;New-LinodeVPC&#8221; command:</strong></p>



<pre class="wp-block-code"><code>PS /&gt; Get-Help New-LinodeVPC

NAME
    New-LinodeVPC

SYNTAX
    New-LinodeVPC &#91;&#91;-description] &lt;string&gt;] &#91;&#91;-apiVersion] &lt;string&gt;] &#91;&#91;-label] &lt;string&gt;] &#91;&#91;-token] &lt;string&gt;] &#91;&#91;-region] &lt;string&gt;] &#91;&#91;-subnetlabel] &lt;string&gt;] &#91;&#91;-iprange] &lt;string&gt;] &#91;&lt;CommonParameters&gt;]

ALIASES
    None

REMARKS
    None</code></pre>



<p></p>



<h4 class="wp-block-heading"></h4>



<p></p>



<h4 class="wp-block-heading">Cool stuff and other quality of life improvements</h4>



<p>Compared to Linode&#8217;s existing CLI, this module will allow you to specify a label for most of the resources you&#8217;re working with. In the background the module will convert the label to the ID of the resource for you.</p>



<p>Example of retrieving an instance details using&nbsp;<a href="https://www.linode.com/docs/products/tools/cli/guides/install/">linode-cli</a>:</p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="415" src="https://blog.slepcevic.net/wp-content/uploads/2024/06/linodecli-1024x415.png" alt="" class="wp-image-517" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/06/linodecli-1024x415.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/06/linodecli-300x121.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/06/linodecli-768x311.png 768w, https://blog.slepcevic.net/wp-content/uploads/2024/06/linodecli.png 1089w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<p><strong>Example retrieving an instance details using Linode Powershell module:</strong></p>



<pre class="wp-block-code"><code>PS C:\&gt; Get-Linodeinstance -label blog.dummyserver.net

id               : 12374123
label            : blog.dummyserver.net
group            :
status           : running
created          : 2024-02-29T17:53:49
updated          : 2024-06-29T00:35:47
type             : g6-standard-2
ipv4             : {172.123.123.123}
ipv6             : 2600:4c0e::f04c:95ff:fe48:cda4/128
image            : linode/ubuntu22.04
region           : nl-ams
specs            : @{disk=81920; memory=4096; vcpus=2; gpus=0; transfer=4000}
alerts           : @{cpu=180; network_in=10; network_out=10; transfer_quota=80; io=10000}
backups          : @{enabled=True; available=True; schedule=; last_successful=2024-06-29T00:29:24}
hypervisor       : kvm
watchdog_enabled : True
tags             : {prod}
host_uuid        : 123f8e495e9d51c53df4fa074e5844bce944a068
has_user_data    : True
placement_group  :
lke_cluster_id   :</code></pre>



<p></p>



<h4 class="wp-block-heading">Development status as of 30.06.2024</h4>



<figure class="wp-block-table alignwide"><table><tbody><tr><td>Percent complete</td><td><strong>51.35%</strong></td></tr><tr><td>Total API commands:</td><td>368</td></tr><tr><td>Implemented Powershell commands</td><td>189</td></tr></tbody></table></figure>



<p>Live status: <a href="https://docs.google.com/spreadsheets/d/1UCzCSL3fRUB6cJKZJBUFEq-hPzaxGcS7B37nP12Ee3I/edit?gid=1065465499#gid=1065465499" target="_blank" rel="noopener" title="">https://docs.google.com/spreadsheets/d/1UCzCSL3fRUB6cJKZJBUFEq-hPzaxGcS7B37nP12Ee3I/edit?gid=1065465499#gid=1065465499</a></p>



<p>Development is focused on providing the <strong>most important features from the start</strong>, so that&#8217;s the reason you will see a lot of less important functionalities being implemented halfway or &#8220;almost there&#8221;.</p>



<p>This is a very active project and my expectation is that in the next few weeks there will be full coverage.</p>



<p>Status of implementation is as follows:</p>



<figure class="wp-block-table"><table><thead><tr><th>Area</th><th>Status</th><th>Notes</th></tr></thead><tbody><tr><td><strong>Account</strong></td><td><strong>Done</strong></td><td>All functionalities besides billing available.</td></tr><tr><td><strong>Beta programs</strong></td><td><strong>Done</strong></td><td><strong>All functionalities available</strong></td></tr><tr><td><strong>Account availability</strong></td><td><strong>Done</strong></td><td><strong>All functionalities available</strong></td></tr><tr><td><strong>Child accounts</strong></td><td>Not started</td><td>None</td></tr><tr><td><strong>Payments</strong></td><td>Not started</td><td>None</td></tr><tr><td><strong>Entity transfers</strong></td><td>Not started</td><td>None</td></tr><tr><td><strong>Events</strong></td><td>WIP</td><td>Everything besides &#8220;Mark as read/seen&#8221; has been implemented</td></tr><tr><td><strong>Invoices</strong></td><td><strong>Done</strong></td><td><strong>All functionalities available</strong></td></tr><tr><td><strong>Logins</strong></td><td><strong>Done</strong></td><td><strong>All functionalities available</strong></td></tr><tr><td><strong>Maintenances</strong></td><td><strong>Done</strong></td><td><strong>All functionalities available</strong></td></tr><tr><td><strong>Notifications</strong></td><td><strong>Done</strong></td><td><strong>All functionalities available</strong></td></tr><tr><td><strong>Oauth Clients</strong></td><td>Not started</td><td>None</td></tr><tr><td><strong>Client thumbnails</strong></td><td>Not started</td><td>None</td></tr><tr><td><strong>Payment methods</strong></td><td>Not started</td><td>None</td></tr><tr><td><strong>Promo credits</strong></td><td>Not started</td><td>None</td></tr><tr><td><strong>Service transfers</strong></td><td>Not started</td><td>None</td></tr><tr><td><strong>Account settings</strong></td><td>Not started</td><td>None</td></tr><tr><td><strong>Settings</strong></td><td>Not started</td><td>None</td></tr><tr><td><strong>Account transfer</strong></td><td><strong>Done</strong></td><td><strong>All functionalities available.</strong></td></tr><tr><td><strong>Users</strong></td><td><strong>Done</strong></td><td>Everything besides &#8220;Update user&#8217;s grant&#8221; command is implemented.</td></tr><tr><td><strong>Beta programs</strong></td><td><strong>Done</strong></td><td><strong>All functionalities available.</strong></td></tr><tr><td><strong>Databases</strong></td><td>WIP</td><td>MySQL 95% done, PostgreSQL 0%.</td></tr><tr><td><strong>Domains</strong></td><td>WIP</td><td>All GET commands done, 80% of write commands done.</td></tr><tr><td><strong>Images</strong></td><td><strong>Done</strong></td><td><strong>All functionalities available.</strong></td></tr><tr><td><strong>Instances</strong></td><td>WIP</td><td>60% done. Most important features like VM management are already available</td></tr><tr><td><strong>StackScripts</strong></td><td>WIP</td><td>80% done. All &#8220;GET&#8221; commands available</td></tr><tr><td><strong>LKE</strong></td><td>WIP</td><td>95% done. All &#8220;GET&#8221; commands available</td></tr><tr><td><strong>Longview</strong></td><td>WIP</td><td>0% done</td></tr><tr><td><strong>Managed</strong></td><td>WIP</td><td>0% done</td></tr><tr><td><strong>Networking</strong></td><td>WIP</td><td>80% done. All &#8220;GET&#8221; commands available</td></tr><tr><td><strong>NodeBalancers</strong></td><td>WIP</td><td>80% done. All &#8220;GET&#8221; commands available</td></tr><tr><td><strong>Object storage</strong></td><td><strong>Done</strong></td><td><strong>All functionalities available.</strong></td></tr><tr><td><strong>Placement groups</strong></td><td>WIP</td><td>0% done</td></tr><tr><td><strong>Profile</strong></td><td>WIP</td><td>0% done</td></tr><tr><td><strong>Regions</strong></td><td><strong>Done</strong></td><td><strong>All functionalities available.</strong></td></tr><tr><td><strong>Support</strong></td><td>WIP</td><td>0% done</td></tr><tr><td><strong>Tags</strong></td><td><strong>Done</strong></td><td><strong>All functionalities available.</strong></td></tr><tr><td><strong>Volumes</strong></td><td>WIP</td><td>50% done</td></tr><tr><td><strong>VPC</strong></td><td><strong>Done</strong></td><td><strong>All functionalities available.</strong></td></tr></tbody></table></figure>



<h4 class="wp-block-heading">Roadmap</h4>



<p>Since this is a &#8220;weekend&#8221; project, no dates will be listed here <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f642.png" alt="🙂" class="wp-smiley" style="height: 1em; max-height: 1em;" /></p>



<p> Project is under very active development and I&#8217;m expecting to finish the first stage in the next few weeks</p>



<p></p>



<ul class="wp-block-list">
<li><strong>Stage 1 &lt; -WE ARE HERE (30.06.2024)</strong>
<ul class="wp-block-list">
<li>Full functionality coverage &#8211; 45% done</li>
</ul>
</li>



<li>Stage 2
<ul class="wp-block-list">
<li>Write Pester tests for all functionalities</li>
</ul>
</li>



<li>Stage 3
<ul class="wp-block-list">
<li>Build test and release pipelines</li>
</ul>
</li>



<li>Stage 4
<ul class="wp-block-list">
<li>Get a signing certificate &amp; publish to NuGet gallery if project gets enough traction</li>
</ul>
</li>
</ul>



<h4 class="wp-block-heading">Contributing</h4>



<p>To report a bug or request a feature in the module, please open a&nbsp;<a href="https://github.com/slepix/LinodePowerShell/issues">GitHub Issue</a>. Will be happy to check it out &lt;3</p>



<p>Have fun with it, please provide feedback (especially around naming commandlets) and until next time <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f600.png" alt="😀" class="wp-smiley" style="height: 1em; max-height: 1em;" /></p>



<p>Alex!</p><p>The post <a href="https://blog.slepcevic.net/powershell-module-for-linode-akamai-connected-cloud/">Powershell module for Linode – Akamai Connected Cloud</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></content:encoded>
					
					<wfw:commentRss>https://blog.slepcevic.net/powershell-module-for-linode-akamai-connected-cloud/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Deploying a GLOBAL sentiment Analysis service using DeepSparse and Akamai Connected Cloud</title>
		<link>https://blog.slepcevic.net/deploying-a-global-sentiment-analysis-service-using-deepsparse-and-akamai-connected-cloud/</link>
					<comments>https://blog.slepcevic.net/deploying-a-global-sentiment-analysis-service-using-deepsparse-and-akamai-connected-cloud/#respond</comments>
		
		<dc:creator><![CDATA[Alesandro Slepčević]]></dc:creator>
		<pubDate>Wed, 12 Jun 2024 22:31:06 +0000</pubDate>
				<category><![CDATA[Akamai Connected Cloud]]></category>
		<category><![CDATA[Akamai Gecko]]></category>
		<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[Cloud]]></category>
		<category><![CDATA[Linode Gecko]]></category>
		<category><![CDATA[Terraform]]></category>
		<category><![CDATA[Akamai CLB]]></category>
		<category><![CDATA[Akamai Cloud Load Balancer]]></category>
		<category><![CDATA[DeepSparse]]></category>
		<category><![CDATA[EdgeAI]]></category>
		<category><![CDATA[Global computing]]></category>
		<category><![CDATA[Infrastructure]]></category>
		<category><![CDATA[Linode VM]]></category>
		<category><![CDATA[Linux]]></category>
		<category><![CDATA[Neural Magic]]></category>
		<guid isPermaLink="false">https://blog.slepcevic.net/?p=441</guid>

					<description><![CDATA[<p>In the previous post, we explored how to deploy a sentiment analysis application using Neural Magic’s DeepSparse on Akamai Connected Cloud (Linode). We leveraged just two dual-core VMs and a Nodebalancer to process a pretty impressive number(40K) of movie reviews...</p>
<p>The post <a href="https://blog.slepcevic.net/deploying-a-global-sentiment-analysis-service-using-deepsparse-and-akamai-connected-cloud/">Deploying a GLOBAL sentiment Analysis service using DeepSparse and Akamai Connected Cloud</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></description>
										<content:encoded><![CDATA[<p>In the <a href="https://blog.slepcevic.net/sentiment-analysis-of-40-thousand-movie-reviews-in-20-minutes-using-neural-magics-deepsparse-inference-runtime-and-linode/" target="_blank" rel="noopener" title="previous post">previous post</a>, we explored how to deploy a sentiment analysis application using Neural Magic’s DeepSparse on Akamai Connected Cloud (Linode). </p>



<p>We leveraged just two dual-core VMs and a Nodebalancer to process a pretty impressive number(40K) of movie reviews in just 20 minutes. However, deploying in a single region can lead to latency/availability issues and doesn&#8217;t fully utilize the global reach of modern cloud infrastructure Akamai Connected Cloud offers. </p>



<p>Also, single region deployments are kinda boring <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f642.png" alt="🙂" class="wp-smiley" style="height: 1em; max-height: 1em;" /></p>



<p>In this post, we&#8217;ll expand our deployment by setting up virtual machines in <strong>all available</strong> Linode regions and replacing the current <a href="https://www.linode.com/products/nodebalancers/" title="Nodebalancer ">Nodebalancer </a>with Akamai’s new <a href="https://www.linode.com/green-light/" title="Cloud Load Balancer">Cloud Load Balancer</a> (currently in beta access). </p>



<p><strong>What is Akamai&#8217;s new Cloud Load Balancer you may ask? It&#8217;s really cool peace of tech.</strong></p>



<p>Think of it like an umbrella over the internet; it gives you the possibility to load balance your workloads across ANY location; it can be on prem, Akamai&#8217;s cloud, some other hyper-scaler, heck, it can even be your home IP address if you want to <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f642.png" alt="🙂" class="wp-smiley" style="height: 1em; max-height: 1em;" /> </p>



<p>As long as it can be reached over the internet, Cloud Load Balancer can use it to deliver the request. </p>



<p>Joking aside, here&#8217;s a more official description of the service:</p>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow">
<p>The Akamai Cloud Load Balancer (ACLB) (formerly referred to as Akamai Global Load Balancer) is a layer 4 and 7 load balancer that distributes traffic based on performance, weight, and content (HTTP headers, query strings, etc.). The ACLB is multi-region, independent of Akamai Delivery, and built for East-West and North-South traffic. Key features include multi-region/multi-cloud load balancing and method selection.</p>
</blockquote>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="259" src="https://blog.slepcevic.net/wp-content/uploads/2024/06/clb3-1024x259.png" alt="" class="wp-image-447" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/06/clb3-1024x259.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/06/clb3-300x76.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/06/clb3-768x195.png 768w, https://blog.slepcevic.net/wp-content/uploads/2024/06/clb3.png 1303w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<h3 class="wp-block-heading">Why Scale Globally?</h3>



<p>Scaling out our application across multiple regions has several benefits:</p>



<ol class="wp-block-list">
<li><strong>Reduced Latency</strong>: By having servers closer to our users, we can significantly reduce the time it takes for requests to travel back and forth.</li>



<li><strong>High Availability</strong>: Distributing the load across multiple regions ensures that if one region goes down, well, we kinda don&#8217;t care, our app stays online <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f642.png" alt="🙂" class="wp-smiley" style="height: 1em; max-height: 1em;" /></li>



<li><strong>Better Performance</strong>: I mean, you can&#8217;t beat physics; simply having the possibility to do compute closer to the user improves performance and user experience.</li>
</ol>



<h3 class="wp-block-heading">Step-by-Step Deployment Guide</h3>



<p>Ok, let&#8217;s get into the meaty part; codebase from our previous post hasn&#8217;t changed significantly; only thing which we changed is that we&#8217;re not hardcoding out region anymore, but we are fetching the list of available regions from Linode API and deploying an instance in each region. </p>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow">
<p>In the code which fetches the regions, you will notice that I commented out the authentication part. </p>



<p>Some regions are available only to authenticated users; if you&#8217;re one of those, just uncomment those few lines and the full region list will be returned to you. </p>
</blockquote>



<p>Let&#8217;s start with terraform plan and see what we will create. </p>



<pre class="wp-block-code"><code>terraform plan</code></pre>



<figure class="wp-block-image size-full"><img loading="lazy" decoding="async" width="942" height="690" src="https://blog.slepcevic.net/wp-content/uploads/2024/06/1.png" alt="" class="wp-image-449" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/06/1.png 942w, https://blog.slepcevic.net/wp-content/uploads/2024/06/1-300x220.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/06/1-768x563.png 768w" sizes="auto, (max-width: 942px) 100vw, 942px" /></figure>



<p>Ok, 25 instances, just what we expect since Akamai has 25 compute regions currently publicly available. </p>



<p>Let&#8217;s proceed with terraform apply </p>



<pre class="wp-block-code"><code>terraform apply</code></pre>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="690" src="https://blog.slepcevic.net/wp-content/uploads/2024/06/2-1024x690.png" alt="" class="wp-image-450" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/06/2-1024x690.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/06/2-300x202.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/06/2-768x518.png 768w, https://blog.slepcevic.net/wp-content/uploads/2024/06/2.png 1300w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow">
<p>I don&#8217;t know about you, but I always nerd out on seeing a bunch of servers popping up in the console <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f600.png" alt="😀" class="wp-smiley" style="height: 1em; max-height: 1em;" /></p>
</blockquote>



<p>In a minute or two we should have all instances deployed. After the instances have been deployed, <strong>cloud-init</strong> will kick off and install DeepSparse server with an Nginx proxy in front (check out the previous post or the YAML file in the repo for more details).</p>



<p><strong>Awesome</strong>! After we&#8217;ve got the infrastructure up and running, last step is to add our nodes to the Cloud load balancer pool; at the moment we will need to do some Clickops <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f641.png" alt="🙁" class="wp-smiley" style="height: 1em; max-height: 1em;" /> CLB service is currently in beta so IaC support isn&#8217;t out yet. </p>



<p>First step is creating a Cloud Load Balancer by clicking the &#8220;<strong>Create Cloud Load Balancer</strong>&#8221; button and giving it a name.</p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="410" src="https://blog.slepcevic.net/wp-content/uploads/2024/06/4-1024x410.png" alt="" class="wp-image-452" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/06/4-1024x410.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/06/4-300x120.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/06/4-768x307.png 768w, https://blog.slepcevic.net/wp-content/uploads/2024/06/4.png 1322w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow">
<p>During the beta period, Cloud Load Balancer is deployed only in 5 locations. This number will grow drastically once the service goes GA. </p>
</blockquote>



<p>With Akamai&#8217;s Cloud Load Balancer, everything starts with a &#8220;<strong>Configuration</strong>&#8220;. Let&#8217;s create one by pressing &#8220;<strong>Add Configuration</strong>&#8221; button. </p>



<p>We will configure our load balancer to &#8220;listen&#8221; on both HTTP and HTTPS. Once we selected HTTPS as our protocol, we need to add a certificate. </p>



<p>In order to do that, we need to prepare our <strong>certificate</strong> and <strong>private key</strong> which we will paste into the configuration field.<em> In this case I will use a self-signed certificate. </em></p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1298" height="578" src="https://blog.slepcevic.net/wp-content/uploads/2024/06/6-1024x456.png" alt="" class="wp-image-453" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/06/6-1024x456.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/06/6-300x134.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/06/6-768x342.png 768w, https://blog.slepcevic.net/wp-content/uploads/2024/06/6.png 1298w" sizes="auto, (max-width: 1298px) 100vw, 1298px" /></figure>



<p>At this stage we will only cover the configuration for the HTTPS protocol, HTTP is really easy and won&#8217;t bother wasting your time on it. </p>


<div class="wp-block-image">
<figure class="aligncenter size-full"><img loading="lazy" decoding="async" width="485" height="886" src="https://blog.slepcevic.net/wp-content/uploads/2024/06/7.png" alt="" class="wp-image-454" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/06/7.png 485w, https://blog.slepcevic.net/wp-content/uploads/2024/06/7-164x300.png 164w" sizes="auto, (max-width: 485px) 100vw, 485px" /></figure>
</div>


<p>We need to paste in the certificate &amp; the key, enter the SNI hostname and press &#8220;<strong>Create and Add</strong>&#8221; button. </p>



<p>After we&#8217;ve got the configuration and the certificate added, we need to create a &#8220;<strong>Route</strong>&#8220;. Let&#8217;s click on  &#8220;<strong>Create a New HTTP Route</strong>&#8221; button and give it a name. </p>


<div class="wp-block-image">
<figure class="aligncenter size-full"><img loading="lazy" decoding="async" width="485" height="483" src="https://blog.slepcevic.net/wp-content/uploads/2024/06/9.png" alt="" class="wp-image-455" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/06/9.png 485w, https://blog.slepcevic.net/wp-content/uploads/2024/06/9-300x300.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/06/9-150x150.png 150w" sizes="auto, (max-width: 485px) 100vw, 485px" /></figure>
</div>


<p>Great, we&#8217;ve created a route, but the route is currently empty and it doesn&#8217;t route anything. We will come back to this a bit later. </p>



<p>Next step is to save our configuration and click on &#8220;<strong>Service Targets</strong>&#8221; tab. </p>



<p>This is the place where we will define our target groups and origin servers. Click on &#8220;<strong>New Service Target</strong>&#8221; button</p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="349" src="https://blog.slepcevic.net/wp-content/uploads/2024/06/10-1024x349.png" alt="" class="wp-image-456" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/06/10-1024x349.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/06/10-300x102.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/06/10-768x262.png 768w, https://blog.slepcevic.net/wp-content/uploads/2024/06/10.png 1301w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<p>Next steps are quite explanatory, we need to give it a name and add the nodes which we want to load balance across. </p>



<p><strong>Remember, this can be one of the existing Linode instance or it can be ANY IP address which can be reached via the internet. </strong></p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="440" src="https://blog.slepcevic.net/wp-content/uploads/2024/06/11-1024x440.png" alt="" class="wp-image-457" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/06/11-1024x440.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/06/11-300x129.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/06/11-768x330.png 768w, https://blog.slepcevic.net/wp-content/uploads/2024/06/11-1536x660.png 1536w, https://blog.slepcevic.net/wp-content/uploads/2024/06/11.png 1801w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<p>This is the step where I could really use IaC support, we need to add all 25 servers by using ClickOps to our &#8220;<strong>Endpoints</strong>&#8221; list.  <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f642.png" alt="🙂" class="wp-smiley" style="height: 1em; max-height: 1em;" /></p>



<p>This is also the place where you can also select the <strong>load balancing algorithm</strong> which will be used to balance requests between the nodes. At the moment there are 5 of them available:</p>



<ul class="wp-block-list">
<li><strong>Round Robin</strong></li>



<li><strong>Least Request</strong></li>



<li><strong>Ring Hash</strong></li>



<li><strong>Random</strong></li>



<li><strong>Maglev</strong></li>
</ul>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="547" src="https://blog.slepcevic.net/wp-content/uploads/2024/06/15-1024x547.png" alt="" class="wp-image-460" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/06/15-1024x547.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/06/15-300x160.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/06/15-768x410.png 768w, https://blog.slepcevic.net/wp-content/uploads/2024/06/15-1536x821.png 1536w, https://blog.slepcevic.net/wp-content/uploads/2024/06/15.png 1783w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<p>Last step in &#8220;<strong>Service Target</strong>&#8221; configuration is to set the path and host header we will use for the health checks on the nodes and click on &#8220;<strong>Save Service Target</strong>&#8221; button. </p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="508" src="https://blog.slepcevic.net/wp-content/uploads/2024/06/12-1024x508.png" alt="" class="wp-image-458" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/06/12-1024x508.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/06/12-300x149.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/06/12-768x381.png 768w, https://blog.slepcevic.net/wp-content/uploads/2024/06/12-1536x762.png 1536w, https://blog.slepcevic.net/wp-content/uploads/2024/06/12.png 1807w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<p><strong>We&#8217;re almost there, I promise <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f600.png" alt="😀" class="wp-smiley" style="height: 1em; max-height: 1em;" /> </strong></p>



<p>Final step is to go back to the &#8220;<strong>Routes</strong>&#8221; tab, click on the route which we&#8217;ve created earlier and click on Edit button.</p>



<p>In the rule configuration we will enter the hostname which we want to match upon and select our &#8220;<strong>Service target</strong>&#8221; from the dropdown. </p>



<p>We can also do advanced request matching based on <strong>path, header, method, regex or query string</strong> but for now we will use path prefix. </p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="535" src="https://blog.slepcevic.net/wp-content/uploads/2024/06/13-1-1024x535.png" alt="" class="wp-image-461" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/06/13-1-1024x535.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/06/13-1-300x157.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/06/13-1-768x401.png 768w, https://blog.slepcevic.net/wp-content/uploads/2024/06/13-1-1536x803.png 1536w, https://blog.slepcevic.net/wp-content/uploads/2024/06/13-1.png 1814w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<p>Ok, we have configured our Cloud Load balancer; final step to get our application running is to create a CNAME record for my subdomain &#8220;<strong>clbtest.slepcevic.net</strong>&#8221; and point it to &#8220;<strong>MDEE053110.mesh.akadns.net</strong>&#8221; (<em>visible in the Summary page of the load balancer</em>). </p>



<p>Let&#8217;s go ahead and visit our website. We see that our DeepSparse API server is happily responding and ready to receive requests! Woohoo!</p>



<p><strong>Yes, it&#8217;s that easy. </strong>In less than 10 minutes we have deployed a globally distributed application on Akamai Connected Cloud. Once IaC support for Cloud Load Balancer is rolled out, we can bring this time down to 5 minutes without any problems. </p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="649" src="https://blog.slepcevic.net/wp-content/uploads/2024/06/16-1024x649.png" alt="" class="wp-image-462" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/06/16-1024x649.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/06/16-300x190.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/06/16-768x487.png 768w, https://blog.slepcevic.net/wp-content/uploads/2024/06/16.png 1264w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<h2 class="wp-block-heading">Ok, 25 regions is cool, but that isn&#8217;t truly global is it? <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f642.png" alt="🙂" class="wp-smiley" style="height: 1em; max-height: 1em;" /></h2>



<p>Yes, you&#8217;re right; with 25 regions we have covered the large majority of the global (internet) population. Can we do better? For sure! Welcome Gecko!</p>



<h3 class="wp-block-heading">Gecko?</h3>



<p>Akamai’s new initiative, code-named Gecko, is set to revolutionize cloud computing by integrating cloud capabilities directly into Akamai&#8217;s extensive edge network. This move aligns perfectly with Akamai’s strategy to provide high-performance, low-latency, and globally scalable solutions. By embedding compute capabilities at or VERY near the edge, Gecko aims to deliver workloads closer to users, devices, and data sources than ever before.</p>



<h3 class="wp-block-heading">What Does Gecko Mean for Our Deployment in the future?</h3>



<p>Gecko&#8217;s will enable us to deploy our sentiment analysis application in hundreds of new locations worldwide, <strong>including traditionally hard-to-reach areas</strong>. This means extremely reduced latency, improved performance, and enhanced availability for users across the world. </p>



<h3 class="wp-block-heading">The Benefits of Deploying on Gecko</h3>



<ol class="wp-block-list">
<li><strong>Ultra-Low Latency</strong>: By running our workloads even closer to end-users, we can drastically reduce the time it takes to process and respond to requests.</li>



<li><strong>Global Reach</strong>: With Gecko, we can deploy in cities and regions where traditional cloud providers struggle to reach, ensuring a truly global presence. </li>



<li><strong>Scalability and Flexibility</strong>: With Akamai&#8217;s large compute footprint, we can scale out our application to tens of thousands of nodes across hundreds of locations. </li>



<li><strong>Consistent Experience</strong>: Let&#8217;s be real, if you&#8217;re running a global application, you&#8217;re most probably dealing with multiple providers; with Gecko we can consolidate all of your workloads and location coverage with a single provider. Just the operational benefits of that should be enough to &#8220;tickle&#8221; your brain into considering it for your application. </li>
</ol>



<h2 class="wp-block-heading">Want to try it yourself?</h2>



<p>Have fun <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f642.png" alt="🙂" class="wp-smiley" style="height: 1em; max-height: 1em;" /> Go ahead and clone the repo from <a href="https://github.com/slepix/neuralmagic-globalLinode" target="_blank" rel="noopener" title="">https://github.com/slepix/neuralmagic-globalLinode</a> and sign up to receive beta access to the new Cloud Load balancer on <a href="https://www.linode.com/green-light/" target="_blank" rel="noopener" title="">https://www.linode.com/green-light/</a> .  </p>



<p>NeuralMagic running across all Linode (including Gecko) regions in the next post? Perhaps <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f642.png" alt="🙂" class="wp-smiley" style="height: 1em; max-height: 1em;" /></p>



<p>Cheers! Alex</p><p>The post <a href="https://blog.slepcevic.net/deploying-a-global-sentiment-analysis-service-using-deepsparse-and-akamai-connected-cloud/">Deploying a GLOBAL sentiment Analysis service using DeepSparse and Akamai Connected Cloud</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></content:encoded>
					
					<wfw:commentRss>https://blog.slepcevic.net/deploying-a-global-sentiment-analysis-service-using-deepsparse-and-akamai-connected-cloud/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Easy way to deploy Linode instances across all (core) regions using Terraform</title>
		<link>https://blog.slepcevic.net/easy-way-to-deploy-linode-instances-across-all-core-regions-using-terraform/</link>
					<comments>https://blog.slepcevic.net/easy-way-to-deploy-linode-instances-across-all-core-regions-using-terraform/#respond</comments>
		
		<dc:creator><![CDATA[Alesandro Slepčević]]></dc:creator>
		<pubDate>Mon, 03 Jun 2024 22:57:35 +0000</pubDate>
				<category><![CDATA[Akamai Connected Cloud]]></category>
		<category><![CDATA[Cloud]]></category>
		<category><![CDATA[Linux]]></category>
		<category><![CDATA[Open Source]]></category>
		<category><![CDATA[Terraform]]></category>
		<category><![CDATA[Infrastructure]]></category>
		<category><![CDATA[Linode regions]]></category>
		<guid isPermaLink="false">https://blog.slepcevic.net/?p=426</guid>

					<description><![CDATA[<p>In preparations for a upcoming blog post related to running AI workloads on Akamai Connected Cloud, I needed to deploy instances in all available Linode regions. Since there&#8217;s no way that I&#8217;ll manually create a list of regions, I thought...</p>
<p>The post <a href="https://blog.slepcevic.net/easy-way-to-deploy-linode-instances-across-all-core-regions-using-terraform/">Easy way to deploy Linode instances across all (core) regions using Terraform</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></description>
										<content:encoded><![CDATA[<p>In preparations for a upcoming blog post related to running AI workloads on Akamai Connected Cloud, I needed to deploy instances in all available Linode regions. Since there&#8217;s no way that I&#8217;ll manually create a list of regions, I thought it would be cool to use Terraforms HTTP provider to dynamically fetch the list of regions and deploy an instance to it. </p>



<p></p>



<h3 class="wp-block-heading">What’s the Plan?</h3>



<p></p>



<ol class="wp-block-list">
<li><strong>Fetch the list of regions</strong>: Use Terraform&#8217;s HTTP provider to retrieve the list of Linode regions via the Linode API.</li>



<li><strong>Deploy Instances</strong>: Use Terraform’s Linode provider to deploy instances in each of these regions.</li>



<li><strong>Grab some coffee</strong>: Sit back and relax while Terraform does the heavy lifting.</li>
</ol>



<h3 class="wp-block-heading">Pre-requisites</h3>



<p>Before we begin, make sure you have:</p>



<ul class="wp-block-list">
<li>A Linode account (If not, sign up <a href="https://www.linode.com/" target="_blank" rel="noopener" title="">here</a>).</li>



<li>A Linode API token (Grab it from your <a href="https://cloud.linode.com/profile/tokens" target="_blank" rel="noopener" title="">Linode dashboard</a>).</li>



<li>Terraform installed on your machine (Download it from <a href="https://developer.hashicorp.com/terraform/install" target="_blank" rel="noopener" title="">here</a>).</li>
</ul>



<h3 class="wp-block-heading">Step-by-Step Guide</h3>



<h4 class="wp-block-heading">Step 1: Define Providers and Variables</h4>



<p>We’ll start by defining our providers and variables. The HTTP provider will fetch the regions, and the Linode provider will deploy our instances.</p>



<p></p>



<pre class="wp-block-code"><code>provider "http" {}

variable "token" {
type = string
default = "mytoken" # Replace with your actual Linode API token
}</code></pre>



<h4 class="wp-block-heading">Step 2: Fetch Regions from Linode API.</h4>



<p>Next, let’s fetch the list of regions using the HTTP provider. We’ll decode the JSON response to extract the region IDs. </p>



<p><strong>Please note that if you uncomment the &#8220;request_headers&#8221; part, you will get a list of regions only available to your user/account. </strong> By default you will get a list of all public regions. </p>



<pre class="wp-block-code"><code>data "http" "regions" {
  url = "https://api.linode.com/v4/regions"

#   request_headers = {
#     Authorization = "Bearer ${var.token}"
#   }
}

locals {
  regions = &#91;for region in jsondecode(data.http.regions.response_body).data : region.id]
}</code></pre>



<h4 class="wp-block-heading">Step 3: Deploy Linode Instances</h4>



<p>Now comes the fun part! We’ll utilize Terraform’s &#8220;<code>for_each"</code> feature to loop through the regions and deploy a virtual machine. </p>



<pre class="wp-block-code"><code><code>provider "linode" {
  token = var.token
}

resource "linode_instance" "instances" {
  for_each = toset(local.regions)

  label    = "linode-${each.key}"
  region   = each.key
  type     = "g6-standard-1"  # Example Linode type, adjust as needed
  image    = "linode/ubuntu20.04"  # Example image, adjust as needed
  root_pass = "your_secure_password"  # Replace with a secure password
  authorized_keys = &#91;file("~/.ssh/id_rsa.pub")]  # Adjust path to your public key
}<span style="background-color: initial; font-family: inherit; font-size: inherit; color: initial;"></span></code></code></pre>



<h4 class="wp-block-heading">Step 4: Outputs</h4>



<p>Finally, let’s define some outputs to see the regions and instances we’ve created.</p>



<pre class="wp-block-code"><code><code>output "regions" {
  value = local.regions
}

output "instances" {
  value = { for instance in linode_instance.instances : instance.id => instance }
}
</code></code></pre>



<h3 class="wp-block-heading">Wrapping Up</h3>



<p>And there you have it! With just a few lines of code, we’ve automated the deployment of Linode instances across all regions. Terraform takes care of the heavy lifting, allowing you to focus on what matters most – enjoying a cup of coffee while your infrastructure magically sets itself up :D+</p>



<figure class="wp-block-image size-full"><img loading="lazy" decoding="async" width="974" height="441" src="http://blog.slepcevic.net/wp-content/uploads/2024/06/Screenshot-2024-06-04-004302.png" alt="" class="wp-image-434" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/06/Screenshot-2024-06-04-004302.png 974w, https://blog.slepcevic.net/wp-content/uploads/2024/06/Screenshot-2024-06-04-004302-300x136.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/06/Screenshot-2024-06-04-004302-768x348.png 768w" sizes="auto, (max-width: 974px) 100vw, 974px" /></figure>



<p>Until next time, Alex!</p><p>The post <a href="https://blog.slepcevic.net/easy-way-to-deploy-linode-instances-across-all-core-regions-using-terraform/">Easy way to deploy Linode instances across all (core) regions using Terraform</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></content:encoded>
					
					<wfw:commentRss>https://blog.slepcevic.net/easy-way-to-deploy-linode-instances-across-all-core-regions-using-terraform/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Sentiment analysis of 40 thousand movie reviews in 20 minutes using Neural Magic&#8217;s DeepSparse inference runtime and Linode virtual machines.</title>
		<link>https://blog.slepcevic.net/sentiment-analysis-of-40-thousand-movie-reviews-in-20-minutes-using-neural-magics-deepsparse-inference-runtime-and-linode/</link>
					<comments>https://blog.slepcevic.net/sentiment-analysis-of-40-thousand-movie-reviews-in-20-minutes-using-neural-magics-deepsparse-inference-runtime-and-linode/#respond</comments>
		
		<dc:creator><![CDATA[Alesandro Slepčević]]></dc:creator>
		<pubDate>Sat, 30 Mar 2024 22:54:00 +0000</pubDate>
				<category><![CDATA[Akamai Connected Cloud]]></category>
		<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[Cloud]]></category>
		<category><![CDATA[Linux]]></category>
		<category><![CDATA[Open Source]]></category>
		<category><![CDATA[Terraform]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[DeepSparse]]></category>
		<category><![CDATA[Linode]]></category>
		<category><![CDATA[Neural Magic]]></category>
		<guid isPermaLink="false">https://blog.slepcevic.net/?p=287</guid>

					<description><![CDATA[<p>First, let me start with a word or two about DeepSparse. DeepSparse is a sparsity-aware inference runtime that delivers GPU-class performance on commodity CPUs, purely in software, anywhere. GPUs Are Not Optimal &#8211; Machine learning inference has evolved over the...</p>
<p>The post <a href="https://blog.slepcevic.net/sentiment-analysis-of-40-thousand-movie-reviews-in-20-minutes-using-neural-magics-deepsparse-inference-runtime-and-linode/">Sentiment analysis of 40 thousand movie reviews in 20 minutes using Neural Magic’s DeepSparse inference runtime and Linode virtual machines.</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></description>
										<content:encoded><![CDATA[<p>First, let me start with a word or two about DeepSparse. </p>



<p><strong>DeepSparse is a sparsity-aware inference runtime that delivers GPU-class performance on commodity CPUs, purely in software, anywhere.</strong></p>



<p><strong>GPUs Are Not Optimal</strong> &#8211; Machine learning inference has evolved over the years led by GPU advancements. GPUs are fast and powerful, but they can be expensive, have shorter life spans, and require a lot of electricity and cooling.</p>



<p>Other major problems with GPU&#8217;s, especially if you&#8217;re thinking in the context of <a href="https://www.akamai.com/newsroom/press-release/akamai-takes-cloud-computing-to-the-edge" title="">Edge computing</a>, is that they can&#8217;t be packed as densely and are power ineffective compared to CPU&#8217;s; not to mention availability these days.</p>



<p>Since Akamai recently partnered up with Neural Magic, I&#8217;ve decided to write a quick tutorial on how to easily get started with running a simple <strong>DeepSparse sentiment analysis workload</strong>. </p>



<p>In case you want more about Akamai and Neural Magic&#8217;s partnership, make sure to watch this excellent video from TFiR. It will also give you a great summary of Akamai&#8217;s Project Gecko.</p>



<figure class="wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio"><div class="wp-block-embed__wrapper">
<div class="responsive-embed widescreen"><iframe loading="lazy" title="Akamai partners with Neural Magic to bring AI to edge use cases" width="1000" height="563" src="https://www.youtube.com/embed/MG45UM4SlbQ?feature=oembed" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" referrerpolicy="strict-origin-when-cross-origin" allowfullscreen></iframe></div>
</div></figure>



<p></p>



<h3 class="wp-block-heading">What is Sentiment analysis?</h3>



<p><strong>Sentiment analysis</strong>&nbsp;(also known as&nbsp;<strong>opinion mining</strong>&nbsp;or&nbsp;<strong>emotion AI</strong>) is the use of&nbsp;<a href="https://en.wikipedia.org/wiki/Natural_language_processing">natural language processing</a>,&nbsp;<a href="https://en.wikipedia.org/wiki/Text_analytics">text analysis</a>,&nbsp;<a href="https://en.wikipedia.org/wiki/Computational_linguistics">computational linguistics</a>, and&nbsp;<a href="https://en.wikipedia.org/wiki/Biometrics">biometrics</a>&nbsp;to systematically identify, extract, quantify, and study affective states and subjective information. Sentiment analysis is widely applied to&nbsp;<a href="https://en.wikipedia.org/wiki/Voice_of_the_customer">voice of the customer</a>&nbsp;materials such as reviews and survey responses, online and social media, and healthcare materials for applications that range from&nbsp;<a href="https://en.wikipedia.org/wiki/Marketing">marketing</a>&nbsp;to&nbsp;<a href="https://en.wikipedia.org/wiki/Customer_relationship_management">customer service</a>&nbsp;to clinical medicine.&nbsp;</p>



<p>Why is DeepSparse cool? Because I&#8217;m doing analysis of 40 thousands movie reviews in 20 minutes using only <strong>TWO DUAL CORE Linode VM&#8217;s. Mind officially blown. </strong></p>



<p></p>



<p></p>



<p>Let&#8217;s do some math here; rounding it up to 120 thousand processed reviews an hour, with 2 instances and a load balancer, we can process over<strong> 86 million requests a month</strong> which will cost you a <strong>staggering 82$ <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f600.png" alt="😀" class="wp-smiley" style="height: 1em; max-height: 1em;" /></strong>. </p>



<p><strong>If you&#8217;re doing that on other cloud providers, you&#8217;re paying a five digit monthly bill for that pleasure. </strong></p>



<p></p>



<h2 class="wp-block-heading">Want to try it yourself? It&#8217;s easy!</h2>



<p>If you want to try it out on Linode, follow instructions below. </p>



<p>If you want to check out Neural Magic DeepSparse repo, head out <a href="https://github.com/neuralmagic/deepsparse" target="_blank" rel="noopener" title="">here</a>.</p>



<p><strong>Step 1.  Clone the Repository</strong>.</p>



<p>Open your terminal or command prompt and run the following command: </p>



<pre class="wp-block-code"><code>git clone https://github.com/slepix/neuralmagic-linode</code></pre>



<p>This code will deploy <strong>2 x Dedicated 4 GB</strong> virtual machines and a <strong>Nodebalancer</strong>. It will also install Neural Magic&#8217;s DeepSparse runtime as a Linux service and  install &amp; configure Nginx to proxy requests to DeepSparse server listening on 127.0.0.1:5543. </p>



<p class="has-vivid-red-color has-text-color has-link-color wp-elements-f0a0be9819cd2bc070b1912e9e812e46"><strong>WARNING: THIS IS NOT PRODUCTION GRADE SERVER CONFIGURATION!</strong></p>



<p class="has-vivid-red-color has-text-color has-link-color wp-elements-42cceec5227b00bdbbe795cf30c2587a">It&#8217;s just a POC! Secure your servers and consult Neural Magic documentation if you want to go to production. </p>



<p><strong>Step 2. </strong>&#8211; <strong>Terraform init</strong></p>



<p>Navigate to the repo using the following command: </p>



<pre class="wp-block-code"><code>cd neuralmagic-linode</code></pre>



<p>If you haven&#8217;t already installed Terraform on your machine, you can download it from the <a href="https://developer.hashicorp.com/terraform/install?product_intent=terraform" target="_blank" rel="noopener" title="">official Terraform website</a> and follow the installation instructions for your operating system.</p>



<p><strong>Step 3. </strong></p>



<p>Initialize Terraform by running:</p>



<pre class="wp-block-code"><code>terraform init</code></pre>



<p><strong>Step 4. </strong>&#8211; <strong>Configure your Linode token</strong></p>



<p>Open <strong>variables.tf </strong>file and paste in your Linode token. If you don&#8217;t know how to create a Linode PAT, check this article <strong><a href="https://www.linode.com/docs/products/tools/api/guides/manage-api-tokens/" target="_blank" rel="noopener" title="">here</a></strong>. It should look similar like the picture. You can also adjust the region while you&#8217;re here <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f600.png" alt="😀" class="wp-smiley" style="height: 1em; max-height: 1em;" /></p>



<pre class="wp-block-preformatted">Token in the picture is not valid. It's just an example. </pre>



<figure class="wp-block-image size-full"><img loading="lazy" decoding="async" width="772" height="275" src="https://blog.slepcevic.net/wp-content/uploads/2024/03/tokendemo.png" alt="" class="wp-image-293" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/03/tokendemo.png 772w, https://blog.slepcevic.net/wp-content/uploads/2024/03/tokendemo-300x107.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/03/tokendemo-768x274.png 768w" sizes="auto, (max-width: 772px) 100vw, 772px" /></figure>



<p><strong>Step 5</strong> &#8211; <strong>Run Terraform apply</strong></p>



<p>After configuring your variables, you can apply the Terraform configuration by running:</p>



<pre class="wp-block-code"><code>terraform apply</code></pre>



<p>Terraform will show you a plan of the changes it intends to make. </p>



<p>Review the plan carefully, and if everything looks good, type &#8220;<code><strong>yes"</strong></code> and press Enter to apply the changes. Give it 5-6 minutes to finish everything and by visiting your Nodebalancer IP, you should be presented with a landing page for DeepSparse server API. </p>



<figure class="wp-block-image size-large"><img loading="lazy" decoding="async" width="1024" height="518" src="https://blog.slepcevic.net/wp-content/uploads/2024/03/Screenshot-2024-03-31-233941-1024x518.png" alt="" class="wp-image-294" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/03/Screenshot-2024-03-31-233941-1024x518.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/03/Screenshot-2024-03-31-233941-300x152.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/03/Screenshot-2024-03-31-233941-768x388.png 768w, https://blog.slepcevic.net/wp-content/uploads/2024/03/Screenshot-2024-03-31-233941.png 1457w" sizes="auto, (max-width: 1024px) 100vw, 1024px" /></figure>



<p><strong>Step 6. </strong></p>



<p>After the installation is done, it&#8217;s finally time to send some data to our API and see how it performs. </p>



<p>We can do that by using <strong>curl </strong>or <strong>invoke-webrequest</strong> if you&#8217;re on Windows and using Powershell. </p>



<p><strong>CURL: </strong></p>



<pre class="wp-block-code"><code>sentence="Neural Magic &amp; Akamai are cool!"
nodebalancer="172.233.34.110" #PUT YOUR NODEBALANCER IP HERE
curl -X POST http://$nodebalancer/v2/models/sentiment_analysis/infer -H "Content-Type: application/json" -d "{\"sequences\": \"$sentence\"}"</code></pre>



<p><strong>PowerShell:</strong></p>



<p></p>



<pre class="wp-block-code"><code>$sentence = "Neural Magic &amp; Akamai are cool!"
$nodebalancer = "172.233.34.110"

$path = "v2/models/sentiment_analysis/infer"
$api = "http://$nodebalancer/$path"
$body = @{
   sequences = $sentence
} | ConvertTo-Json

(Invoke-WebRequest -Uri $api -Method Post -ContentType "application/json" -Body $body -ErrorAction Stop).content</code></pre>



<p>In both cases make sure to paste in the <strong>IP address of the Nodebalancer</strong> you deployed and modify the sentence as you wish. </p>



<h2 class="wp-block-heading">Benchmark time!</h2>



<p>In the repository, I&#8217;ve included a file called movies.csv and three files; two PowerShell and one Python file.</p>



<p><strong>movies.zip</strong> &#8211; unzip this one in the same folder where your benchmark scripts are. </p>



<p><strong>analyze.ps1</strong> &#8211; PowerShell based benchmark, sends requests in serial &#8211; not performant. </p>



<p><strong>panalyze.ps1</strong> &#8211; PowerShell based benchmark, sends requests in parallel &#8211; better performant</p>



<p><strong>pypanalyze.py</strong> &#8211; Python based benchmark, sends requests in parallel &#8211; <strong>best performer (doh!) &lt;-use this</strong></p>



<figure class="wp-block-image size-full"><img loading="lazy" decoding="async" width="581" height="153" src="http://blog.slepcevic.net/wp-content/uploads/2024/04/Screenshot-2024-04-01-014021.png" alt="" class="wp-image-327" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/04/Screenshot-2024-04-01-014021.png 581w, https://blog.slepcevic.net/wp-content/uploads/2024/04/Screenshot-2024-04-01-014021-300x79.png 300w" sizes="auto, (max-width: 581px) 100vw, 581px" /></figure>



<p>All you need to do to in order to kick off a benchmark is to update the the URL variable with your Nodebalancer IP and you&#8217;re off to the races. </p>



<h2 class="wp-block-heading">Does it scale?</h2>



<p><strong>Yes!</strong> For kicks I&#8217;ve added a third node and the same job finished in 825 seconds. Feel free to add as many nodes as you like and see what numbers you can get. Additionally, you can play with the number of workers in the Python file. </p>



<pre class="wp-block-preformatted">Note 1: python script has been written with the help of ChatGPT :) Results matched with my PowerShell version against verified smaller sample size(check note 2), so I'm gonna call it good :)
	 
Note 2: PowerShell versions don't handle some comments as they should and end up sending garbage to the API. Happens in 3% of the cases. Most probably some encoding/character issue which I couldn't be bothered to fix :)

Note3: Movies.csv file has been generated by using data from https://kaggle.com/

</pre>



<p>Cheers, </p>



<p>Alex. </p>



<p></p><p>The post <a href="https://blog.slepcevic.net/sentiment-analysis-of-40-thousand-movie-reviews-in-20-minutes-using-neural-magics-deepsparse-inference-runtime-and-linode/">Sentiment analysis of 40 thousand movie reviews in 20 minutes using Neural Magic’s DeepSparse inference runtime and Linode virtual machines.</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></content:encoded>
					
					<wfw:commentRss>https://blog.slepcevic.net/sentiment-analysis-of-40-thousand-movie-reviews-in-20-minutes-using-neural-magics-deepsparse-inference-runtime-and-linode/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Deploying HarperDB Docker container on Linode VM using Terraform and cloud-init</title>
		<link>https://blog.slepcevic.net/deploying-harperdb-docker-container-on-linode-vm-using-terraform-and-cloud-init/</link>
					<comments>https://blog.slepcevic.net/deploying-harperdb-docker-container-on-linode-vm-using-terraform-and-cloud-init/#respond</comments>
		
		<dc:creator><![CDATA[Alesandro Slepčević]]></dc:creator>
		<pubDate>Tue, 27 Feb 2024 12:45:04 +0000</pubDate>
				<category><![CDATA[Akamai Connected Cloud]]></category>
		<category><![CDATA[Cloud]]></category>
		<category><![CDATA[Database]]></category>
		<category><![CDATA[Linux]]></category>
		<category><![CDATA[Terraform]]></category>
		<category><![CDATA[HarperDB]]></category>
		<category><![CDATA[Linode]]></category>
		<guid isPermaLink="false">http://172.233.40.105/blog.slepcevic.net/?p=227</guid>

					<description><![CDATA[<p>This blog post is a small update to the excellent post guys from HarperDB wrote &#8211; https://www.harperdb.io/development/tutorials/deploying-harperdb-on-digital-ocean-linode-with-terraform Since Linode now support cloud-init and metadata service, I decided to extend their example by using cloud-init to do the installation of Docker...</p>
<p>The post <a href="https://blog.slepcevic.net/deploying-harperdb-docker-container-on-linode-vm-using-terraform-and-cloud-init/">Deploying HarperDB Docker container on Linode VM using Terraform and cloud-init</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></description>
										<content:encoded><![CDATA[<p>This blog post is a small update to the excellent post guys from HarperDB wrote &#8211; <a href="https://www.harperdb.io/development/tutorials/deploying-harperdb-on-digital-ocean-linode-with-terraform" target="_blank" rel="noreferrer noopener">https://www.harperdb.io/development/tutorials/deploying-harperdb-on-digital-ocean-linode-with-terraform</a></p>



<p>Since Linode now support <a href="https://www.linode.com/docs/guides/applications/configuration-management/cloud-init/">cloud-init</a> and <a href="https://www.linode.com/docs/guides/using-metadata-cloud-init-on-any-distribution/">metadata service</a>, I decided to extend their example by using cloud-init to do the installation of Docker Engine and HarperDB container. </p>



<p>All you need to do to get HarperDB running is to copy all of these files in the same folder while making sure to keep the filenames the same (<strong>harperdb.yaml</strong> at least). </p>



<p>After that, simply run Terraform init, then Terraform apply and in 2-3 minutes you should have your HarperDB instance up and running. </p>



<figure class="wp-block-image"><img loading="lazy" decoding="async" width="1083" height="540" src="https://blog.slepcevic.net/wp-content/uploads/2024/02/harpervscode.png" alt="This is how your folder structure should look like when you copy all the files. " class="wp-image-253" srcset="https://blog.slepcevic.net/wp-content/uploads/2024/02/harpervscode.png 1083w, https://blog.slepcevic.net/wp-content/uploads/2024/02/harpervscode-300x150.png 300w, https://blog.slepcevic.net/wp-content/uploads/2024/02/harpervscode-1024x511.png 1024w, https://blog.slepcevic.net/wp-content/uploads/2024/02/harpervscode-768x383.png 768w" sizes="auto, (max-width: 1083px) 100vw, 1083px" /></figure>



<figure class="wp-block-pullquote has-small-font-size" style="font-style:normal;font-weight:700"><blockquote><p><strong>Please change the passwords for your Linode VM and HarperDB in the compute.tf &amp; harperdb.yaml files &lt;3</strong><br><strong> Don&#8217;t be that guy! <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f600.png" alt="😀" class="wp-smiley" style="height: 1em; max-height: 1em;" /></strong></p><cite>Alex, 2024</cite></blockquote></figure>



<pre class="wp-block-code"><code>terraform init
terraform apply -var="token=YourLinodeToken"</code></pre>



<p><strong>harperdb.yaml</strong></p>



<pre class="wp-block-code"><code>#cloud-config
runcmd:
  - sudo apt-get update
  - sudo apt-get install -yq ca-certificates curl gnupg lsb-release
  - sudo install -m 0755 -d /etc/apt/keyrings
  - for pkg in docker.io docker-doc docker-compose docker-compose-v2 podman-docker containerd runc; do sudo apt-get remove $pkg; done
  - curl -fsSL https://download.docker.com/linux/ubuntu/gpg | sudo gpg --dearmor -o /etc/apt/keyrings/docker.gpg
  - sudo chmod a+r /etc/apt/keyrings/docker.gpg
  - echo "deb &#091;arch=$(dpkg --print-architecture) signed-by=/etc/apt/keyrings/docker.gpg] https://download.docker.com/linux/ubuntu $(. /etc/os-release &amp;&amp; echo "$VERSION_CODENAME") stable" | sudo tee /etc/apt/sources.list.d/docker.list &gt; /dev/null
  - sudo apt-get update
  - sudo apt-get install -y docker-ce docker-ce-cli containerd.io docker-buildx-plugin docker-compose-plugin
  - sudo mkdir /home/harperdb
  - sudo chmod 777 /home/harperdb
  - sudo docker run -d -v /home/harperdb:/home/harperdb/hdb -e HDB_ADMIN_USERNAME=HDB_ADMIN -e HDB_ADMIN_PASSWORD=<strong>password </strong>-p 9925:9925 -p 9926:9926 harperdb/harperdb</code></pre>



<p><strong>compute.tf</strong></p>



<pre class="wp-block-code"><code>resource "linode_instance" "harperdb" {
  image = "linode/ubuntu22.04"
  region = "nl-ams" #Pick The region you want
  type = "g6-standard-1"
  root_pass = "YourRootPassword!" #Change this :D
   metadata {
    user_data = base64encode(file("${path.module}/harperdb.yaml"))
  }
}

resource "linode_firewall" "harperdb_firewall" {
  label = "harperdb"

  inbound {
    label    = "ssh"
    action   = "ACCEPT"
    protocol = "TCP"
    ports    = "22"
    ipv4     = &#091;"0.0.0.0/0"]
    ipv6     = &#091;"::/0"]
  }

  inbound {
    label    = "harperdb"
    action   = "ACCEPT"
    protocol = "TCP"
    ports    = "9925-9926"
    ipv4     = &#091;"0.0.0.0/0"]
    ipv6     = &#091;"::/0"]
  }

  inbound_policy = "DROP"
  outbound_policy = "ACCEPT"
  linodes = &#091;linode_instance.harperdb.id]
}

terraform {
  required_providers {
    linode = {
      source = "linode/linode"
    }
  }
}

provider "linode" {
  token = var.token
}

variable "token" {
    default = ""
    type = string
}</code></pre>



<p>Alex. </p><p>The post <a href="https://blog.slepcevic.net/deploying-harperdb-docker-container-on-linode-vm-using-terraform-and-cloud-init/">Deploying HarperDB Docker container on Linode VM using Terraform and cloud-init</a> first appeared on <a href="https://blog.slepcevic.net">Architect the cloud</a>.</p>]]></content:encoded>
					
					<wfw:commentRss>https://blog.slepcevic.net/deploying-harperdb-docker-container-on-linode-vm-using-terraform-and-cloud-init/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
	</channel>
</rss>

<!--
Performance optimized by W3 Total Cache. Learn more: https://www.boldgrid.com/w3-total-cache/?utm_source=w3tc&utm_medium=footer_comment&utm_campaign=free_plugin

Page Caching using Disk: Enhanced 
Lazy Loading (feed)

Served from: blog.slepcevic.net @ 2026-01-01 04:11:57 by W3 Total Cache
-->