<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Javascript on Bil Arikan</title><link>https://bil.arikan.ca/tags/javascript/</link><description>Recent content in Javascript on Bil Arikan</description><generator>Hugo</generator><language>en-US</language><lastBuildDate>Mon, 06 Apr 2026 00:00:00 -0500</lastBuildDate><atom:link href="https://bil.arikan.ca/tags/javascript/index.xml" rel="self" type="application/rss+xml"/><item><title>In-app Live Assistant : Part 2 --- Building the Screen-Sharing Version</title><link>https://bil.arikan.ca/posts/in-app-live-assistant-part-2-building-the/</link><pubDate>Mon, 06 Apr 2026 00:00:00 -0500</pubDate><guid>https://bil.arikan.ca/posts/in-app-live-assistant-part-2-building-the/</guid><description>The ADK Dev UI supports camera but not screen sharing. In this post I build a custom client that swaps getUserMedia for getDisplayMedia, writes a 16kHz AudioWorklet from scratch, and sends 1 FPS JPEG screen snapshots to the same WebSocket endpoint. The agent now sees the application instead of the user&amp;rsquo;s face.</description></item></channel></rss>