<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>FastTextContext on Volundarhus</title><link>http://www.volundarhus.com/categories/fasttextcontext/</link><description>Recent content in FastTextContext on Volundarhus</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Sun, 05 Apr 2026 19:52:47 -0700</lastBuildDate><atom:link href="http://www.volundarhus.com/categories/fasttextcontext/index.xml" rel="self" type="application/rss+xml"/><item><title>Stratified Word Embeddings with Patient and Provider Metadata</title><link>http://www.volundarhus.com/posts/fast_text_context/</link><pubDate>Sun, 05 Apr 2026 19:52:47 -0700</pubDate><guid>http://www.volundarhus.com/posts/fast_text_context/</guid><description>&lt;p&gt;Standard word embeddings treat every occurrence of a word as equivalent. &amp;ldquo;Chest pain&amp;rdquo; means the same thing whether it appears in a note written by an emergency attending about an 80-year-old Medicare patient or by a resident describing an elective pre-op. That homogeneity is a fundamental design choice in Word2Vec and FastText, and for most downstream tasks it is a reasonable one. Clinical NLP, however, operates in a domain where who is speaking, about whom, and in what care setting are inseparable from meaning. This post describes &lt;strong&gt;FastTextContext&lt;/strong&gt;, a from-scratch C++ implementation that extends FastText&amp;rsquo;s skip-gram model with learned metadata embeddings for patient demographics and provider role, fused through a shared projection matrix.&lt;/p&gt;</description></item></channel></rss>