chore: post

This commit is contained in:
Cory Dransfeldt 2024-05-08 12:37:27 -07:00
parent 21c39c2ee6
commit 58c1147e83
No known key found for this signature in database
6 changed files with 372 additions and 6 deletions

View file

@ -33,12 +33,10 @@ export default async (request) => {
.single()
if (albumError && albumError.code === 'PGRST116') {
const albumImageUrl = `https://coryd.dev/media/albums/${albumKey}.jpg`
const albumMBID = null
const { error: insertAlbumError } = await supabase.from('albums').insert([
{
mbid: albumMBID,
image: albumImageUrl,
mbid: null,
image: `https://coryd.dev/media/albums/${albumKey}.jpg`,
key: albumKey,
name: album,
tentative: true
@ -54,6 +52,32 @@ export default async (request) => {
return new Response(JSON.stringify({ status: 'error', message: albumError.message }), { headers: { "Content-Type": "application/json" } })
}
const { data: artistData, error: artistError } = await supabase
.from('artists')
.select('*')
.eq('key', artistKey)
.single()
if (artistError && artistError.code === 'PGRST116') {
const { error: insertArtistError } = await supabase.from('artists').insert([
{
mbid: null,
image: `https://coryd.dev/media/artists/${artistKey}.jpg`,
key: albumKey,
name: album,
tentative: true
}
])
if (insertArtistError) {
console.error('Error inserting artist into Supabase:', insertArtistError.message)
return new Response(JSON.stringify({ status: 'error', message: insertArtistError.message }), { headers: { "Content-Type": "application/json" } })
}
} else if (artistError) {
console.error('Error querying artist from Supabase:', artistError.message)
return new Response(JSON.stringify({ status: 'error', message: artistError.message }), { headers: { "Content-Type": "application/json" } })
}
const { error: listenError } = await supabase.from('listens').insert([
{
artist_name: artist,

View file

@ -34,9 +34,11 @@ export default {
macos: '#macOS #Apple',
mastodon: '#Mastodon',
music: '#Music',
musicbrainz: '#MusicBrainz',
mystery: '#Mystery',
netlify: '#Netlify',
nonfiction: '#NonFiction',
plex: '#Plex',
politics: '#Politics',
privacy: '#Privacy',
productivity: '#Productivity',
@ -49,6 +51,7 @@ export default {
'social media': '#SocialMedia',
sports: '#Sports',
spotify: '#Music',
supabase: '#Supabase',
'surveillance capitalism': '#SurveillanceCapitalism',
'tattoos': '#Tattoos',
tech: '#Tech',

View file

@ -4,7 +4,7 @@
{% capture alt %}{{ item.title | escape }} by {{ item.artist }}{% endcapture %}
<div class="item">
<div class="meta">
<img src="https://coryd.dev/.netlify/images/?url={{ item.image }}&fit=cover&w=64&h=64&fm=webp&q=65" class="image-banner" alt="{{ alt }}" loading="lazy" decoding="async" width="64" height="64" />
<img src="https://coryd.dev/.netlify/images/?url={{ item.image }}&fit=cover&w=64&h=64&fm=webp&q=65" alt="{{ alt }}" loading="lazy" decoding="async" width="64" height="64" />
<div class="meta-text">
<div class="title">{{ item.title }}</div>
<div class="subtext">

Binary file not shown.

After

Width:  |  Height:  |  Size: 32 KiB

View file

@ -9,7 +9,7 @@ I've been using Fastmail for years now and have explored a number of different a
For now, I've approached filtering my mail by applying regular expressions to reasonably broad categories of incoming mail[^2]. My thinking with this approach is that will scale better over the long term by applying heuristics to common phrases and patterns in incoming mail without the need to apply rules to senders on a per address or domain basis.
<img src="https://coryd.dev/.netlify/images/?url=https://coryd.dev/media/blog/fastmail-workflow.jpg&w=1000&fm=webp&q=65" class="image-banner" alt="A diagram of my Fastmail workflow" />
<img src="https://coryd.dev/.netlify/images/?url=https://coryd.dev/media/blog/fastmail-workflow.jpg&w=1000&fm=webp&q=65" class="image-banner" loading="eager" decoding="async" alt="A diagram of my Fastmail workflow" width="1000" height="733" />
## Alias-specific rules

View file

@ -0,0 +1,339 @@
---
date: 2024-05-08T11:48-08:00
title: Improving my self-hosted scrobbling implementation
description: I wrote (fairly) recently about implementing my own scrobbler using Plex webhooks, edge functions and blob storage. So far — so far — this has worked quite well.
tags:
- music
- plex
- tech
- supabase
- musicbrainz
---
[I wrote (fairly) recently about implementing my own scrobbler using Plex webhooks, edge functions and blob storage.](https://coryd.dev/posts/2024/building-a-scrobbler-using-plex-webhooks-edge-functions-and-blob-storage/) So far — so far — this has worked quite well.<!-- excerpt -->
In doing this, I was keeping listens as JSON blobs with Netlify, stored under keys for the week they belonged to.
Artist and album metadata lived in similar blob structures. What I found, however, was that this — especially the metadata — quickly became unwieldy. Chalk this up to my inclination to force what I knew to work instead of picking a better tool.
In the interest of making something I've enjoyed building and using more durable, I started looking at flexible database solutions. I poked around, asked some friends and landed on [Supabase](https://supabase.com):
> Supabase is an open source Firebase alternative.
Start your project with a Postgres database, Authentication, instant APIs, Edge Functions, Realtime subscriptions, Storage, and Vector embeddings.
Sounds good, right? (Very much so, learning curve aside.)
I set up three tables: listens, albums and artists.
I had data for each, structured as JSON. I wrote some ugly node scripts (I'll spare you the pain of seeing those here) that wrote these out to CSVs.
I imported those CSVs into their respective tables, and worked my way to connections between the tables that look like this:
<img src="https://coryd.dev/.netlify/images/?url=https://coryd.dev/media/blog/supabase-schema.png&w=1000&fm=webp&q=65" class="image-banner" loading="eager" decoding="async" alt="A diagram of my scrobbling tables" width="1000" height="886" />
The connections between the tables allow me to query data specific to a given listen's artist or album — data is stored in a given table where it makes the most sense: artist `mbid`s with artists, `genre`s with artists and so forth. I can then retrieve that data, provided I have a valid listen, using Supabase's select syntax: `artists (mbid, image)` or `albums (mbid, image)`.
Plex's webhooks send out a fair bit of data, but most of that data is Plex-specific. What I'm able to use from that payload is the art, album and track names. These are stored in the listens table with a timestamp derived when I receive a webhook payload containing a `media.scrobble` event. These webhooks are sent to a `scrobble` endpoint (read: edge function:
```javascript
import { createClient } from '@supabase/supabase-js'
import { DateTime } from 'luxon'
const SUPABASE_URL = Netlify.env.get('SUPABASE_URL')
const SUPABASE_KEY = Netlify.env.get('SUPABASE_KEY')
const supabase = createClient(SUPABASE_URL, SUPABASE_KEY)
const sanitizeMediaString = (string) => string.normalize('NFD').replace(/[\u0300-\u036f\u2010—\.\?\(\)\[\]\{\}]/g, '').replace(/\.{3}/g, '')
export default async (request) => {
const ACCOUNT_ID_PLEX = process.env.ACCOUNT_ID_PLEX
const params = new URL(request.url).searchParams
const id = params.get('id')
if (!id) return new Response(JSON.stringify({ status: 'Bad request' }), { headers: { "Content-Type": "application/json" } })
if (id !== ACCOUNT_ID_PLEX) return new Response(JSON.stringify({ status: 'Forbidden' }), { headers: { "Content-Type": "application/json" } })
const data = await request.formData()
const payload = JSON.parse(data.get('payload'))
if (payload?.event === 'media.scrobble') {
const artist = payload['Metadata']['grandparentTitle']
const album = payload['Metadata']['parentTitle']
const track = payload['Metadata']['title']
const listenedAt = Math.floor(DateTime.now().toSeconds())
const artistKey = sanitizeMediaString(artist).replace(/\s+/g, '-').toLowerCase()
const albumKey = `${artistKey}-${sanitizeMediaString(album).replace(/\s+/g, '-').toLowerCase()}`
const { data: albumData, error: albumError } = await supabase
.from('albums')
.select('*')
.eq('key', albumKey)
.single()
if (albumError && albumError.code === 'PGRST116') {
const { error: insertAlbumError } = await supabase.from('albums').insert([
{
mbid: null,
image: `https://coryd.dev/media/albums/${albumKey}.jpg`,
key: albumKey,
name: album,
tentative: true
}
])
if (insertAlbumError) {
console.error('Error inserting album into Supabase:', insertAlbumError.message)
return new Response(JSON.stringify({ status: 'error', message: insertAlbumError.message }), { headers: { "Content-Type": "application/json" } })
}
} else if (albumError) {
console.error('Error querying album from Supabase:', albumError.message)
return new Response(JSON.stringify({ status: 'error', message: albumError.message }), { headers: { "Content-Type": "application/json" } })
}
const { data: artistData, error: artistError } = await supabase
.from('artists')
.select('*')
.eq('key', artistKey)
.single()
if (artistError && artistError.code === 'PGRST116') {
const { error: insertArtistError } = await supabase.from('artists').insert([
{
mbid: null,
image: `https://coryd.dev/media/artists/${artistKey}.jpg`,
key: albumKey,
name: album,
tentative: true
}
])
if (insertArtistError) {
console.error('Error inserting artist into Supabase:', insertArtistError.message)
return new Response(JSON.stringify({ status: 'error', message: insertArtistError.message }), { headers: { "Content-Type": "application/json" } })
}
} else if (artistError) {
console.error('Error querying artist from Supabase:', artistError.message)
return new Response(JSON.stringify({ status: 'error', message: artistError.message }), { headers: { "Content-Type": "application/json" } })
}
const { error: listenError } = await supabase.from('listens').insert([
{
artist_name: artist,
album_name: album,
track_name: track,
listened_at: listenedAt,
album_key: albumKey
}
])
if (listenError) {
console.error('Error inserting data into Supabase:', listenError.message)
console.log('Track with the error:', {
artist_name: artist,
album_name: album,
track_name: track,
listened_at: listenedAt,
album_key: albumKey
})
return new Response(JSON.stringify({ status: 'error', message: listenError.message }), { headers: { "Content-Type": "application/json" } })
}
}
return new Response(JSON.stringify({ status: 'success' }), { headers: { "Content-Type": "application/json" } })
}
export const config = {
path: '/api/scrobble',
}
```
Scary! (Not really). This does a whole host of things:
1. Creates a `supabase` SDK client with my database url and API key.
2. Defines a `sanitizeMediaString` helper that does its best to normalize media strings.
3. Restricts access to requests with my Plex account ID and rejects invalid requests.
4. It then retrieves the data we're interested in from the webhook payload.
5. Next, in the interest of being cautious, we look to see if our `albumKey` is in the albums table. If it's not, we add a record for the album with the `boolean` `tenative` value set to true.
1. This allows me to quickly identify records that were added programmatically and tidy them up if need be.
2. If we can't add a record for the album, we error out as we don't have an album to attribute the listen to.
6. Rinse and repeat for artists (tracks need artists!).
7. Finally, we record the listen — if there's an error we log out which track caused the error.
Whew! On to the big scary (not!) presentation part:
```javascript
import { createClient } from '@supabase/supabase-js'
import { DateTime } from 'luxon'
const SUPABASE_URL = process.env.SUPABASE_URL
const SUPABASE_KEY = process.env.SUPABASE_KEY
const supabase = createClient(SUPABASE_URL, SUPABASE_KEY)
const fetchDataForPeriod = async (startPeriod, fields, table) => {
const PAGE_SIZE = 1000
let rows = []
let rangeStart = 0
while (true) {
const { data, error } = await supabase
.from(table)
.select(fields)
.order('listened_at', { ascending: false })
.gte('listened_at', startPeriod.toSeconds())
.range(rangeStart, rangeStart + PAGE_SIZE - 1)
if (error) {
console.error(error)
break
}
rows = rows.concat(data)
if (data.length < PAGE_SIZE) break
rangeStart += PAGE_SIZE
}
return rows
}
const aggregateData = (data, groupByField, groupByType) => {
const aggregation = {}
data.forEach(item => {
const key = item[groupByField]
if (!aggregation[key]) {
if (groupByType === 'track') {
aggregation[key] = {
title: item[groupByField],
plays: 0,
mbid: item['albums']?.mbid || '',
url: item['albums']?.mbid ? `https://musicbrainz.org/release/${item['albums'].mbid}` : `https://musicbrainz.org/search?query=${encodeURIComponent(item['album_name'])}&type=release`,
image: item['albums']?.image || '',
timestamp: item['listened_at'],
type: groupByType
}
} else {
aggregation[key] = {
title: item[groupByField],
plays: 0,
mbid: item[groupByType]?.mbid || '',
url: item[groupByType]?.mbid ? `https://musicbrainz.org/${groupByType === 'albums' ? 'release' : 'artist'}/${item[groupByType].mbid}` : `https://musicbrainz.org/search?query=${encodeURIComponent(item[groupByField])}&type=${groupByType === 'albums' ? 'release' : 'artist'}`,
image: item[groupByType]?.image || '',
type: groupByType
}
}
if (
groupByType === 'track' ||
groupByType === 'albums'
) aggregation[key]['artist'] = item['artist_name']
}
aggregation[key].plays++
})
return Object.values(aggregation).sort((a, b) => b.plays - a.plays)
}
export default async function() {
const periods = {
week: DateTime.now().minus({ days: 7 }).startOf('day'), // Last week
month: DateTime.now().minus({ days: 30 }).startOf('day'), // Last 30 days
threeMonth: DateTime.now().minus({ months: 3 }).startOf('day'), // Last three months
year: DateTime.now().minus({ years: 1 }).startOf('day'), // Last 365 days
}
const results = {}
const selectFields = `
track_name,
artist_name,
album_name,
album_key,
listened_at,
artists (mbid, image),
albums (mbid, image)
`
for (const [period, startPeriod] of Object.entries(periods)) {
const periodData = await fetchDataForPeriod(startPeriod, selectFields, 'listens')
results[period] = {
artists: aggregateData(periodData, 'artist_name', 'artists'),
albums: aggregateData(periodData, 'album_name', 'albums'),
tracks: aggregateData(periodData, 'track_name', 'track')
}
}
const recentData = await fetchDataForPeriod(DateTime.now().minus({ days: 7 }), selectFields, 'listens')
results.recent = {
artists: aggregateData(recentData, 'artist_name', 'artists'),
albums: aggregateData(recentData, 'album_name', 'albums'),
tracks: aggregateData(recentData, 'track_name', 'track')
}
results.nowPlaying = results.recent.tracks[0]
return results
}
```
1. We set up another `supabase` SDK client with my database url and API key.
2. We author a method that will fetch data for different time periods (cleverly named `fetchDataForPeriod`). We pass in a start period, it pages through results from our `listens` table until we're out and filters out anything with a `listend_at` timestamp greater than or equal to our `startPeriod` in seconds.
3. This all yields data that we store in `periodData` — which is called within a for loop where we pull data for each of our time periods defined in `const periods = ...`.
1. Within this loop we populate our `results` object. It has a key for each period we're concerned with populating and that key is populated with aggregated `artists`, `albums` and `tracks` data from our `aggregateData` function (clever naming again, right?).
2. This `aggregateData` function takes our raw data, a field to group by that corresponds to the name of the pertinent column in Supabase and a type.
1. If we're populating track objects, we define one shape, if we're dealing with albums or artists, we modify the shape a bit.
2. If an object is either a track or an album, we add the artist name to the object.
3. If an object has already been built, we increment the play count and go on our merry way.
4. I guess that's a lot — or a bit, I dunno. JSON to CSVs to tables, wire the tables together, write to the table, query things and present it.
This also required that I update my `now-playing` endpoint/edge function that displays the last played track on my home page. The relevant section of that file is here:
```javascript
import { createClient } from '@supabase/supabase-js';
const SUPABASE_URL = process.env.SUPABASE_URL
const SUPABASE_KEY = process.env.SUPABASE_KEY
const supabase = createClient(SUPABASE_URL, SUPABASE_KEY);
...
export default async () => {
const { data, error } = await supabase
.from('listens')
.select(`
track_name,
artist_name,
listened_at,
artists (mbid, genre)
`)
.order('listened_at', { ascending: false })
.range(0, 1)
const headers = {
"Content-Type": "application/json",
"Cache-Control": "public, s-maxage=360, stale-while-revalidate=1080",
};
if (error) {
console.error('Error fetching data:', error);
return new Response(JSON.stringify({ error: "Failed to fetch the latest track" }), { headers });
}
if (data.length === 0) {
return new Response(JSON.stringify({ message: "No recent tracks found" }), { headers });
}
const scrobbleData = data[0]
return new Response(JSON.stringify({
content: `${emojiMap(
scrobbleData.artists.genre,
scrobbleData.artist_name
)} ${scrobbleData.track_name} by <a href="http://musicbrainz.org/artist/${scrobbleData.artists.mbid}">${
scrobbleData.artist_name
}</a>`,
}), { headers });
};
```
1. You guessed it! We set up a `supabase` SDK client.
2. We query our listens table and grab our `mbid` and `genre` via our foreign key connection to the artists table.
3. In much the same way as we did before, we pass the artist `genre` and `name` through our absurd `emojiMap` function and send back HTML to display.
All of the music data used to present artist and album grids and track charts is now retrieved from Supabase on each build. It's much easier to add, modify or deal with artist and album metadata (and even update listen data if needed).