Compare commits
4 Commits
5e8f7e6005
...
47a520e945
Author | SHA1 | Date | |
---|---|---|---|
47a520e945 | |||
f5aceea6fa | |||
3c44fcc062 | |||
5df94eaafa |
1
.gitignore
vendored
1
.gitignore
vendored
@ -23,3 +23,4 @@ yarn-error.log
|
||||
|
||||
# Platform specific
|
||||
.DS_Store
|
||||
*compose-prod.yml
|
||||
|
@ -1,4 +1,4 @@
|
||||
FROM node:20.12.2-alpine3.18 AS base
|
||||
FROM node:22-alpine AS base
|
||||
|
||||
# All deps stage
|
||||
FROM base AS deps
|
||||
@ -6,11 +6,10 @@ WORKDIR /app
|
||||
ADD package.json package-lock.json ./
|
||||
RUN npm ci
|
||||
|
||||
FROM node:22 AS dev-deps
|
||||
FROM deps AS develop
|
||||
WORKDIR /app
|
||||
ADD package.json package-lock.json ./
|
||||
COPY --from=deps /app/node_modules /app/node_modules
|
||||
ENV NODE_ENV=development
|
||||
RUN npm ci
|
||||
EXPOSE 3333
|
||||
|
||||
# Production only deps stage
|
||||
|
@ -53,7 +53,7 @@ export default defineConfig({
|
||||
() => import('@adonisjs/lucid/database_provider'),
|
||||
() => import('@adonisjs/auth/auth_provider'),
|
||||
() => import('@adonisjs/inertia/inertia_provider'),
|
||||
() => import('@adonisjs/redis/redis_provider')
|
||||
() => import('@adonisjs/redis/redis_provider'),
|
||||
],
|
||||
|
||||
/*
|
||||
|
52
app/Helpers/Replays.ts
Normal file
52
app/Helpers/Replays.ts
Normal file
@ -0,0 +1,52 @@
|
||||
import Replay from '#models/replay'
|
||||
import { parseSentryLinkHeader, SentryPagination } from './Sentry.js'
|
||||
|
||||
import env from '#start/env'
|
||||
let recordsUpdated = 0
|
||||
const SENTRY_TOKEN = env.get('SENTRY_TOKEN')
|
||||
interface ApiResponse<T> {
|
||||
data: T
|
||||
// optionally, you can define `meta`, `errors`, etc. if your API returns them
|
||||
}
|
||||
export async function fetchBatch(url: string) {
|
||||
const options: RequestInit = {
|
||||
headers: {
|
||||
Authorization: `Bearer ${SENTRY_TOKEN}`,
|
||||
},
|
||||
}
|
||||
const req = await fetch(url, options)
|
||||
if (!req.ok) {
|
||||
throw new Error(`Request failed with status ${req.status}`)
|
||||
}
|
||||
|
||||
const resp = (await req.json()) as ApiResponse<Replay[]>
|
||||
const replays = resp.data
|
||||
const headers = req.headers
|
||||
|
||||
const cleanedData = replays.map((record) => sanitizeInput(record, Replay.allowedFields))
|
||||
|
||||
let updated = await Replay.updateOrCreateMany('id', cleanedData)
|
||||
recordsUpdated = recordsUpdated + updated.length
|
||||
const linkHeader = headers.get('link')
|
||||
if (!linkHeader) {
|
||||
return { error: 'link header missing from Sentry API response' }
|
||||
}
|
||||
const pagination: SentryPagination = parseSentryLinkHeader(linkHeader)
|
||||
|
||||
if (pagination.hasNextResults == true) {
|
||||
console.log('fetching', pagination.next)
|
||||
await fetchBatch(pagination.next)
|
||||
}
|
||||
console.log('no more results')
|
||||
return { recordsUpdated }
|
||||
}
|
||||
|
||||
function sanitizeInput(data: Record<string, any>, allowedFields: string[]) {
|
||||
return allowedFields.reduce(
|
||||
(acc, key) => {
|
||||
if (key in data) acc[key] = data[key]
|
||||
return acc
|
||||
},
|
||||
{} as Record<string, any>
|
||||
)
|
||||
}
|
27
app/Helpers/Sentry.ts
Normal file
27
app/Helpers/Sentry.ts
Normal file
@ -0,0 +1,27 @@
|
||||
export interface SentryPagination {
|
||||
previous: string
|
||||
hasPreviousResults: boolean
|
||||
hasNextResults: boolean
|
||||
next: string
|
||||
}
|
||||
export function parseSentryLinkHeader(header: string): SentryPagination {
|
||||
const links = header.split(',').map((part) => part.trim())
|
||||
|
||||
let result = {} as SentryPagination
|
||||
for (const link of links) {
|
||||
const match = link.match(/<([^>]+)>;\s*rel="([^"]+)";\s*results="([^"]+)";\s*cursor="([^"]+)"/)
|
||||
if (!match) continue
|
||||
|
||||
const [, url, rel, results] = match
|
||||
|
||||
if (rel === 'previous') {
|
||||
result.previous = url
|
||||
result.hasPreviousResults = results === 'true'
|
||||
} else if (rel === 'next') {
|
||||
result.next = url
|
||||
result.hasNextResults = results === 'true'
|
||||
}
|
||||
}
|
||||
|
||||
return result
|
||||
}
|
21
app/Helpers/Webhook.ts
Normal file
21
app/Helpers/Webhook.ts
Normal file
@ -0,0 +1,21 @@
|
||||
import env from '#start/env'
|
||||
|
||||
export async function sendDataToWebhook(responseData: {
|
||||
version: number
|
||||
updatedAt: Date
|
||||
numberOfRecords: number
|
||||
data: unknown
|
||||
}) {
|
||||
try {
|
||||
console.log('syncing to webhook')
|
||||
await fetch(env.get('WEBHOOK_URL'), {
|
||||
headers: {
|
||||
'content-type': 'application/json',
|
||||
},
|
||||
method: 'POST',
|
||||
body: JSON.stringify(responseData),
|
||||
})
|
||||
} catch (e) {
|
||||
console.error('error sending webhook data', e)
|
||||
}
|
||||
}
|
@ -1,70 +1,39 @@
|
||||
import Replay from '#models/replay'
|
||||
import env from '#start/env'
|
||||
import type { HttpContext } from '@adonisjs/core/http'
|
||||
import db from '@adonisjs/lucid/services/db'
|
||||
const SENTRY_TOKEN = env.get('SENTRY_TOKEN')
|
||||
const SENTRY_ORG = env.get('SENTRY_ORG')
|
||||
let recordsUpdated = 0
|
||||
import redis from '@adonisjs/redis/services/main'
|
||||
import { fetchBatch } from '../Helpers/Replays.js'
|
||||
import { sendDataToWebhook } from '../Helpers/Webhook.js'
|
||||
|
||||
|
||||
interface ApiResponse<T> {
|
||||
data: T;
|
||||
// optionally, you can define `meta`, `errors`, etc. if your API returns them
|
||||
}
|
||||
|
||||
interface SentryPagination {
|
||||
previous: string;
|
||||
hasPreviousResults: boolean;
|
||||
hasNextResults: boolean;
|
||||
next: string
|
||||
}
|
||||
export default class ReplaysController {
|
||||
|
||||
public async stats({ request, response }: HttpContext) {
|
||||
const {sendToWebhook} = request.qs()
|
||||
const cacheKey = `replays:sync:latest_version`
|
||||
const latestFetchVersion = await redis.get(`replays:fetch:latest_version`)
|
||||
const latestQueryVersion = await redis.get(`replays:stats:latest_version`)
|
||||
if (latestFetchVersion == latestQueryVersion) {
|
||||
const { sendToWebhook } = request.qs()
|
||||
const latestVersion = await redis.get(`replays:stats:latest_version`)
|
||||
let results
|
||||
results = await redis.get(`replays:sync:version:${latestQueryVersion}:results`)
|
||||
if (!results) {
|
||||
console.log('no data in cache, updating')
|
||||
results = await getResults()
|
||||
await redis.set(`replays:sync:version:${latestQueryVersion}:results`, JSON.stringify(results))
|
||||
|
||||
}
|
||||
console.log('resultssdsdfds')
|
||||
return response.json(results)
|
||||
if (!latestVersion) {
|
||||
console.log('Cache miss')
|
||||
results = await Replay.updateReplayStats()
|
||||
} else {
|
||||
let results = await getResults()
|
||||
console.log('results quer', latestQueryVersion)
|
||||
await redis.set(`replays:stats:version:${latestQueryVersion}:results`, JSON.stringify(results))
|
||||
await redis.set(`replays:stats:latest_version`, latestFetchVersion)
|
||||
await redis.set(`replays:fetch:latest_version`, latestFetchVersion)
|
||||
return response.json(results)
|
||||
if (sendToWebhook) {
|
||||
try {
|
||||
console.log('syncing to webhook')
|
||||
await fetch(env.get('WEBHOOK_URL'),
|
||||
{
|
||||
headers:
|
||||
{
|
||||
'content-type': 'application/json'
|
||||
},
|
||||
method: 'POST',
|
||||
body: JSON.stringify(results.rows)
|
||||
}
|
||||
)
|
||||
} catch(e) {
|
||||
console.error('error sending webhook data', e)
|
||||
}
|
||||
console.log('cache hit')
|
||||
let data = await redis.get(`replays:stats:version:${latestVersion}:results`)
|
||||
if (data) {
|
||||
results = JSON.parse(data)
|
||||
}
|
||||
}
|
||||
|
||||
response.json(results.rows)
|
||||
let responseData = {
|
||||
version: results.version,
|
||||
updatedAt: results.updatedAt,
|
||||
numberOfRecords: results.rows.length,
|
||||
data: results.rows,
|
||||
}
|
||||
if (sendToWebhook) {
|
||||
await sendDataToWebhook(responseData)
|
||||
}
|
||||
return response.json(responseData)
|
||||
}
|
||||
|
||||
public async list({ request, inertia }: HttpContext) {
|
||||
const page = request.input('page', 1)
|
||||
const perPage = 20
|
||||
@ -74,7 +43,7 @@ export default class ReplaysController {
|
||||
let paginated, meta, replays
|
||||
|
||||
if (data) {
|
||||
({ paginated, meta, replays } = JSON.parse(data))
|
||||
;({ paginated, meta, replays } = JSON.parse(data))
|
||||
} else {
|
||||
paginated = await Replay.query().paginate(page, perPage)
|
||||
paginated.baseUrl('/list')
|
||||
@ -83,7 +52,7 @@ export default class ReplaysController {
|
||||
|
||||
meta = {
|
||||
...json.meta,
|
||||
links: buildPaginationLinks(json.meta)
|
||||
links: buildPaginationLinks(json.meta),
|
||||
}
|
||||
|
||||
replays = json.data
|
||||
@ -94,114 +63,47 @@ export default class ReplaysController {
|
||||
return inertia.render('Replays/Index', {
|
||||
data: {
|
||||
replays,
|
||||
meta
|
||||
}
|
||||
meta,
|
||||
},
|
||||
})
|
||||
|
||||
|
||||
}
|
||||
|
||||
async index({ request, response }: HttpContext) {
|
||||
const { statsPeriod, start, end } = request.qs()
|
||||
recordsUpdated = 0
|
||||
|
||||
let queryString: string = '?statsPeriod=24h'// Default in case none is provided
|
||||
let queryString: string = '?statsPeriod=24h' // Default in case none is provided
|
||||
if (statsPeriod) {
|
||||
queryString = `?statsPeriod=${statsPeriod}`
|
||||
} else if (start && end) {
|
||||
queryString = `?start=${start}&end=${end}`
|
||||
}
|
||||
const replays = await fetchBatch(`https://sentry.io/api/0/organizations/${SENTRY_ORG}/replays/${queryString}`)
|
||||
let latestVersion = await redis.get(`replays:fetch:latest_version`)
|
||||
if (!latestVersion) {
|
||||
redis.set('replays:fetch:latest_version', 1)
|
||||
} else {
|
||||
redis.set('replays:fetch:latest_version', ++latestVersion)
|
||||
}
|
||||
await fetchBatch(`https://sentry.io/api/0/organizations/${SENTRY_ORG}/replays/${queryString}`)
|
||||
|
||||
return response.json(replays)
|
||||
}
|
||||
let queryResults = await Replay.updateReplayStats()
|
||||
|
||||
return response.json({ version: queryResults.latestVersion, ...queryResults })
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
async function fetchBatch(url: string) {
|
||||
const options: RequestInit = {
|
||||
headers: {
|
||||
Authorization: `Bearer ${SENTRY_TOKEN}`
|
||||
}
|
||||
}
|
||||
const req = await fetch(url, options)
|
||||
if (!req.ok) {
|
||||
throw new Error(`Request failed with status ${req.status}`);
|
||||
}
|
||||
|
||||
const resp = await req.json() as ApiResponse<Replay[]>;
|
||||
const replays = resp.data;
|
||||
const headers = req.headers
|
||||
|
||||
const cleanedData = replays.map(record => sanitizeInput(record, Replay.allowedFields))
|
||||
|
||||
let updated = await Replay.updateOrCreateMany('id', cleanedData)
|
||||
recordsUpdated = recordsUpdated + updated.length
|
||||
const linkHeader = headers.get('link')
|
||||
if (!linkHeader) {
|
||||
return { error: 'link header missing from Sentry API response' }
|
||||
}
|
||||
const pagination: SentryPagination = parseSentryLinkHeader(linkHeader)
|
||||
|
||||
if (pagination.hasNextResults == true) {
|
||||
console.log('fetching', pagination.next)
|
||||
await fetchBatch(pagination.next)
|
||||
}
|
||||
console.log('no more results')
|
||||
return { recordsUpdated }
|
||||
|
||||
}
|
||||
function parseSentryLinkHeader(header: string): SentryPagination {
|
||||
const links = header.split(',').map(part => part.trim())
|
||||
|
||||
let result = {} as SentryPagination
|
||||
for (const link of links) {
|
||||
const match = link.match(/<([^>]+)>;\s*rel="([^"]+)";\s*results="([^"]+)";\s*cursor="([^"]+)"/)
|
||||
if (!match) continue
|
||||
|
||||
const [, url, rel, results] = match
|
||||
|
||||
if (rel === 'previous') {
|
||||
result.previous = url
|
||||
result.hasPreviousResults = results === 'true'
|
||||
} else if (rel === 'next') {
|
||||
result.next = url
|
||||
result.hasNextResults = results === 'true'
|
||||
}
|
||||
}
|
||||
|
||||
return result
|
||||
}
|
||||
|
||||
function sanitizeInput(data: Record<string, any>, allowedFields: string[]) {
|
||||
return allowedFields.reduce((acc, key) => {
|
||||
if (key in data) acc[key] = data[key]
|
||||
return acc
|
||||
}, {} as Record<string, any>)
|
||||
}
|
||||
|
||||
function buildPaginationLinks(meta: { previousPageUrl: string, lastPage: number; currentPage: number; nextPageUrl: string }) {
|
||||
function buildPaginationLinks(meta: {
|
||||
previousPageUrl: string
|
||||
lastPage: number
|
||||
currentPage: number
|
||||
nextPageUrl: string
|
||||
}) {
|
||||
const links = []
|
||||
|
||||
// Previous
|
||||
links.push({
|
||||
url: meta.previousPageUrl,
|
||||
label: '« Prev',
|
||||
active: false
|
||||
active: false,
|
||||
})
|
||||
|
||||
for (let page = 1; page <= meta.lastPage; page++) {
|
||||
links.push({
|
||||
url: `/list?page=${page}`,
|
||||
label: page.toString(),
|
||||
active: page === meta.currentPage
|
||||
active: page === meta.currentPage,
|
||||
})
|
||||
}
|
||||
|
||||
@ -209,62 +111,8 @@ function buildPaginationLinks(meta: { previousPageUrl: string, lastPage: number;
|
||||
links.push({
|
||||
url: meta.nextPageUrl,
|
||||
label: 'Next »',
|
||||
active: false
|
||||
active: false,
|
||||
})
|
||||
|
||||
return links
|
||||
}
|
||||
|
||||
async function getResults(){
|
||||
let results = await db.rawQuery(`
|
||||
SELECT
|
||||
u.display_name,
|
||||
u.sessions,
|
||||
u.total_time_seconds,
|
||||
u.total_time_readable,
|
||||
u.average_session_time_readable,
|
||||
u.average_time_seconds,
|
||||
r.id AS last_session_id,
|
||||
r.finished_at AS last_session_time
|
||||
|
||||
FROM (
|
||||
-- Aggregate sessions in the last 30 days
|
||||
SELECT
|
||||
"user" ->> 'display_name' AS display_name,
|
||||
COUNT(duration) AS sessions,
|
||||
SUM(duration) AS total_time_seconds,
|
||||
AVG(duration) AS average_time_seconds,
|
||||
CONCAT(
|
||||
FLOOR(SUM(duration) / 86400), 'd ',
|
||||
FLOOR(MOD(SUM(duration), 86400) / 3600), 'h ',
|
||||
FLOOR(MOD(SUM(duration), 3600) / 60), 'm'
|
||||
) AS total_time_readable,
|
||||
CONCAT(
|
||||
FLOOR(COUNT(duration) / 86400), 'd ',
|
||||
FLOOR(MOD(COUNT(duration), 86400) / 3600), 'h ',
|
||||
FLOOR(MOD(COUNT(duration), 3600) / 60), 'm'
|
||||
) AS average_session_time_readable
|
||||
FROM
|
||||
replays
|
||||
WHERE
|
||||
finished_at >= NOW() - INTERVAL '30 days'
|
||||
GROUP BY
|
||||
"user" ->> 'display_name'
|
||||
) u
|
||||
|
||||
-- LATERAL JOIN to get latest session (either within 30d or fallback to latest overall)
|
||||
JOIN LATERAL (
|
||||
SELECT id, finished_at
|
||||
FROM replays
|
||||
WHERE "user" ->> 'display_name' = u.display_name
|
||||
ORDER BY
|
||||
CASE WHEN finished_at >= NOW() - INTERVAL '30 days' THEN 0 ELSE 1 END,
|
||||
finished_at DESC
|
||||
LIMIT 1
|
||||
) r ON true
|
||||
|
||||
ORDER BY
|
||||
u.total_time_seconds DESC;`
|
||||
)
|
||||
return results
|
||||
}
|
@ -1,7 +1,66 @@
|
||||
import { DateTime } from 'luxon'
|
||||
import { BaseModel, column } from '@adonisjs/lucid/orm'
|
||||
import db from '@adonisjs/lucid/services/db'
|
||||
import redis from '@adonisjs/redis/services/main'
|
||||
|
||||
export default class Replay extends BaseModel {
|
||||
public static async updateReplayStats() {
|
||||
let results = await db.rawQuery(`
|
||||
SELECT
|
||||
u.display_name,
|
||||
u.sessions,
|
||||
u.total_time_seconds,
|
||||
u.total_time_readable,
|
||||
u.average_session_time_readable,
|
||||
u.average_time_seconds,
|
||||
r.id AS last_session_id,
|
||||
r.finished_at AS last_session_time
|
||||
|
||||
FROM (
|
||||
-- Aggregate sessions in the last 30 days
|
||||
SELECT
|
||||
"user" ->> 'display_name' AS display_name,
|
||||
COUNT(duration) AS sessions,
|
||||
SUM(duration) AS total_time_seconds,
|
||||
AVG(duration) AS average_time_seconds,
|
||||
CONCAT(
|
||||
FLOOR(SUM(duration) / 86400), 'd ',
|
||||
FLOOR(MOD(SUM(duration), 86400) / 3600), 'h ',
|
||||
FLOOR(MOD(SUM(duration), 3600) / 60), 'm'
|
||||
) AS total_time_readable,
|
||||
CONCAT(
|
||||
FLOOR(COUNT(duration) / 86400), 'd ',
|
||||
FLOOR(MOD(COUNT(duration), 86400) / 3600), 'h ',
|
||||
FLOOR(MOD(COUNT(duration), 3600) / 60), 'm'
|
||||
) AS average_session_time_readable
|
||||
FROM
|
||||
replays
|
||||
WHERE
|
||||
finished_at >= NOW() - INTERVAL '30 days'
|
||||
AND "user" ->> 'display_name' LIKE '%@%'
|
||||
GROUP BY
|
||||
"user" ->> 'display_name'
|
||||
) u
|
||||
|
||||
-- LATERAL JOIN to get latest session (either within 30d or fallback to latest overall)
|
||||
JOIN LATERAL (
|
||||
SELECT id, finished_at
|
||||
FROM replays
|
||||
WHERE "user" ->> 'display_name' = u.display_name
|
||||
ORDER BY
|
||||
CASE WHEN finished_at >= NOW() - INTERVAL '30 days' THEN 0 ELSE 1 END,
|
||||
finished_at DESC
|
||||
LIMIT 1
|
||||
) r ON true
|
||||
|
||||
ORDER BY
|
||||
u.total_time_seconds DESC;`)
|
||||
const updatedVersion = await redis.incr('replays:stats:latest_version')
|
||||
results.version = updatedVersion
|
||||
results.updatedAt = Date.now()
|
||||
await redis.set(`replays:stats:version:${updatedVersion}:results`, JSON.stringify(results))
|
||||
return results
|
||||
}
|
||||
@column({ isPrimary: true })
|
||||
declare id: string
|
||||
|
||||
@ -12,14 +71,14 @@ export default class Replay extends BaseModel {
|
||||
prepare: (value) => {
|
||||
// The values from sentry are just arrays so convert them to json
|
||||
return JSON.stringify(value)
|
||||
}
|
||||
},
|
||||
})
|
||||
declare trace_ids: string[]
|
||||
|
||||
@column({
|
||||
prepare: (value) => {
|
||||
return JSON.stringify(value)
|
||||
}
|
||||
},
|
||||
})
|
||||
declare error_ids: string[]
|
||||
|
||||
@ -30,26 +89,22 @@ export default class Replay extends BaseModel {
|
||||
prepare: (value) => {
|
||||
// The values from sentry are just arrays so convert them to json
|
||||
return JSON.stringify(value)
|
||||
}
|
||||
},
|
||||
})
|
||||
declare tags: string[]
|
||||
|
||||
@column()
|
||||
declare user: string[]
|
||||
|
||||
|
||||
@column()
|
||||
declare sdk: any
|
||||
|
||||
|
||||
@column()
|
||||
declare os: any
|
||||
|
||||
|
||||
@column()
|
||||
declare browser: any
|
||||
|
||||
|
||||
@column()
|
||||
declare device: any
|
||||
|
||||
@ -59,21 +114,19 @@ export default class Replay extends BaseModel {
|
||||
@column()
|
||||
declare is_archived: boolean | null
|
||||
|
||||
|
||||
@column({
|
||||
prepare: (value) => {
|
||||
// The values from sentry are just arrays so convert them to json
|
||||
return JSON.stringify(value)
|
||||
}
|
||||
},
|
||||
})
|
||||
declare urls: any
|
||||
|
||||
|
||||
@column({
|
||||
prepare: (value) => {
|
||||
// The values from sentry are just arrays so convert them to json
|
||||
return JSON.stringify(value)
|
||||
}
|
||||
},
|
||||
})
|
||||
declare clicks: any
|
||||
|
||||
@ -92,7 +145,7 @@ export default class Replay extends BaseModel {
|
||||
@column.dateTime()
|
||||
declare finished_at: DateTime | null
|
||||
|
||||
@column.dateTime({serializeAs: 'started_at'})
|
||||
@column.dateTime({ serializeAs: 'started_at' })
|
||||
declare started_at: DateTime | null
|
||||
|
||||
@column()
|
||||
@ -110,12 +163,11 @@ export default class Replay extends BaseModel {
|
||||
@column()
|
||||
declare platform: string | null
|
||||
|
||||
|
||||
@column({
|
||||
prepare: (value) => {
|
||||
// The values from sentry are just arrays so convert them to json
|
||||
return JSON.stringify(value)
|
||||
}
|
||||
},
|
||||
})
|
||||
declare releases: any
|
||||
|
||||
|
@ -1,16 +1,14 @@
|
||||
---
|
||||
name: sentry
|
||||
services:
|
||||
scraper:
|
||||
build:
|
||||
context: .
|
||||
target: dev-deps
|
||||
target: develop
|
||||
env_file: .env.develop
|
||||
volumes:
|
||||
- ./:/app
|
||||
- node_modules:/app/node_modules
|
||||
command: /bin/bash -c "echo 'hello' && node ace migration:run --force && node ace serve --watch"
|
||||
|
||||
command: /bin/sh -c "node ace migration:run --force && node ace serve --watch"
|
||||
depends_on:
|
||||
db:
|
||||
condition: service_healthy
|
77
compose.yml
Normal file
77
compose.yml
Normal file
@ -0,0 +1,77 @@
|
||||
---
|
||||
name: sentry
|
||||
services:
|
||||
reverse-proxy:
|
||||
image: traefik:latest
|
||||
command: --api.insecure=true --providers.docker
|
||||
ports:
|
||||
- 80:80
|
||||
- 8080:8080
|
||||
volumes:
|
||||
- /var/run/docker.sock:/var/run/docker.sock
|
||||
networks:
|
||||
- traefik
|
||||
scraper:
|
||||
labels:
|
||||
- 'traefik.enable=true'
|
||||
- 'traefik.docker.network=sentry_traefik'
|
||||
- 'traefik.http.routers.scraper.rule=Host(`sentry.docker.localhost`)'
|
||||
- 'traefik.http.services.scraper.loadbalancer.server.port=3333'
|
||||
- 'traefik.http.routers.scraper.entrypoints=http'
|
||||
- 'traefik.http.routers.scraper.service=scraper'
|
||||
networks:
|
||||
- traefik
|
||||
- redis
|
||||
- database
|
||||
depends_on:
|
||||
db:
|
||||
condition: service_healthy
|
||||
db:
|
||||
image: postgres:16
|
||||
environment:
|
||||
- POSTGRES_PASSWORD=password
|
||||
healthcheck:
|
||||
test: ['CMD-SHELL', 'pg_isready', '-d', 'postgres']
|
||||
interval: 5s
|
||||
timeout: 60s
|
||||
retries: 5
|
||||
start_period: 5s
|
||||
networks:
|
||||
- database
|
||||
volumes:
|
||||
- pg_data:/var/lib/postgresql/data
|
||||
grafana:
|
||||
image: grafana/grafana:latest
|
||||
labels:
|
||||
- 'traefik.enable=true'
|
||||
- 'traefik.docker.network=sentry_traefik'
|
||||
- 'traefik.http.routers.grafana.rule=Host(`grafana.docker.localhost`)'
|
||||
- 'traefik.http.routers.grafana.entrypoints=http'
|
||||
- 'traefik.http.services.grafana.loadbalancer.server.port=3000'
|
||||
- 'traefik.http.routers.grafana.service=grafana'
|
||||
networks:
|
||||
- traefik
|
||||
- database
|
||||
healthcheck:
|
||||
test:
|
||||
[
|
||||
'CMD-SHELL',
|
||||
'wget --no-verbose --tries=1 --spider http://localhost:3000/api/health || exit 1',
|
||||
]
|
||||
interval: 10s
|
||||
timeout: 30s
|
||||
retries: 5
|
||||
start_period: 30s
|
||||
redis:
|
||||
image: redis:latest
|
||||
networks:
|
||||
- redis
|
||||
networks:
|
||||
traefik:
|
||||
driver: bridge
|
||||
database:
|
||||
driver: bridge
|
||||
redis:
|
||||
driver: bridge
|
||||
volumes:
|
||||
pg_data: {}
|
@ -1,65 +0,0 @@
|
||||
---
|
||||
name: sentry
|
||||
services:
|
||||
reverse-proxy:
|
||||
image: traefik:latest
|
||||
command: --api.insecure=true --providers.docker
|
||||
ports:
|
||||
- 80:80
|
||||
- 8080:8080
|
||||
volumes:
|
||||
- /var/run/docker.sock:/var/run/docker.sock
|
||||
networks:
|
||||
- traefik
|
||||
scraper:
|
||||
labels:
|
||||
- "traefik.enable=true"
|
||||
- "traefik.docker.network=sentry_traefik"
|
||||
- "traefik.http.routers.scraper.rule=Host(`sentry.docker.localhost`)"
|
||||
- "traefik.http.services.scraper.loadbalancer.server.port=3333"
|
||||
- "traefik.http.routers.scraper.entrypoints=http"
|
||||
- "traefik.http.routers.scraper.service=scraper"
|
||||
networks:
|
||||
- traefik
|
||||
- redis
|
||||
- database
|
||||
db:
|
||||
image: postgres:16
|
||||
environment:
|
||||
- POSTGRES_PASSWORD=password
|
||||
healthcheck:
|
||||
test: ["CMD-SHELL", "pg_isready", "-d", "postgres"]
|
||||
interval: 5s
|
||||
timeout: 60s
|
||||
retries: 5
|
||||
start_period: 5s
|
||||
networks:
|
||||
- database
|
||||
volumes:
|
||||
- pg_data:/var/lib/postgresql/data
|
||||
grafana:
|
||||
image: grafana/grafana:latest
|
||||
labels:
|
||||
- "traefik.enable=true"
|
||||
- "traefik.docker.network=sentry_traefik"
|
||||
- "traefik.http.routers.grafana.rule=Host(`grafana.docker.localhost`)"
|
||||
- "traefik.http.routers.grafana.entrypoints=http"
|
||||
- "traefik.http.services.grafana.loadbalancer.server.port=3000"
|
||||
- "traefik.http.routers.grafana.service=grafana"
|
||||
|
||||
networks:
|
||||
- traefik
|
||||
- database
|
||||
redis:
|
||||
image: redis:latest
|
||||
networks:
|
||||
- redis
|
||||
networks:
|
||||
traefik:
|
||||
driver: bridge
|
||||
database:
|
||||
driver: bridge
|
||||
redis:
|
||||
driver: bridge
|
||||
volumes:
|
||||
pg_data: {}
|
@ -4,10 +4,10 @@
|
||||
import '../css/app.css'
|
||||
import { createSSRApp, h } from 'vue'
|
||||
import type { DefineComponent } from 'vue'
|
||||
import { createInertiaApp } from '@inertiajs/vue3'
|
||||
import { createInertiaApp, Link } from '@inertiajs/vue3'
|
||||
import { resolvePageComponent } from '@adonisjs/inertia/helpers'
|
||||
|
||||
const appName = import.meta.env.VITE_APP_NAME || 'AdonisJS'
|
||||
Vue.component('inertia-link', Link)
|
||||
|
||||
createInertiaApp({
|
||||
progress: { color: '#5468FF' },
|
||||
|
@ -9,7 +9,6 @@
|
||||
<th class="p-2">Email</th>
|
||||
<th class="p-2">Date</th>
|
||||
<th class="p-2">Location</th>
|
||||
|
||||
</tr>
|
||||
</thead>
|
||||
<tbody>
|
||||
@ -17,13 +16,22 @@
|
||||
<td class="p-2">{{ replay.id }}</td>
|
||||
<td class="p-2">{{ replay.user.email ?? replay.user.display_name }}</td>
|
||||
<td class="p-2">{{ replay.finished_at }}</td>
|
||||
<td class="p-2">{{ replay.user.geo ? `${replay.user.geo.city} ${replay.user.geo.subdivision}, ${replay.user.geo.region}` : 'unknown' }}</td>
|
||||
<td class="p-2">
|
||||
{{
|
||||
replay.user.geo
|
||||
? `${replay.user.geo.city} ${replay.user.geo.subdivision}, ${replay.user.geo.region}`
|
||||
: 'unknown'
|
||||
}}
|
||||
</td>
|
||||
</tr>
|
||||
</tbody>
|
||||
</table>
|
||||
|
||||
<!-- Pagination -->
|
||||
<div class="mt-4 flex flex-wrap items-center gap-2" v-if="data.meta && data.meta.links && data.meta.links.length > 1">
|
||||
<div
|
||||
class="mt-4 flex flex-wrap items-center gap-2"
|
||||
v-if="data.meta && data.meta.links && data.meta.links.length > 1"
|
||||
>
|
||||
<!-- First -->
|
||||
<Link
|
||||
v-if="firstPageUrl && !isFirstPage"
|
||||
@ -34,11 +42,7 @@
|
||||
</Link>
|
||||
|
||||
<!-- Previous -->
|
||||
<Link
|
||||
v-if="prevPageUrl"
|
||||
:href="prevPageUrl"
|
||||
class="px-3 py-1 border rounded text-sm"
|
||||
>
|
||||
<Link v-if="prevPageUrl" :href="prevPageUrl" class="px-3 py-1 border rounded text-sm">
|
||||
‹ Prev
|
||||
</Link>
|
||||
|
||||
@ -50,7 +54,7 @@
|
||||
class="px-3 py-1 border rounded text-sm"
|
||||
:class="{
|
||||
'font-bold bg-gray-300': link.active,
|
||||
'text-gray-400 cursor-not-allowed': !link.url
|
||||
'text-gray-400 cursor-not-allowed': !link.url,
|
||||
}"
|
||||
>
|
||||
<span v-html="link.label" />
|
||||
@ -58,11 +62,7 @@
|
||||
</template>
|
||||
|
||||
<!-- Next -->
|
||||
<Link
|
||||
v-if="nextPageUrl"
|
||||
:href="nextPageUrl"
|
||||
class="px-3 py-1 border rounded text-sm"
|
||||
>
|
||||
<Link v-if="nextPageUrl" :href="nextPageUrl" class="px-3 py-1 border rounded text-sm">
|
||||
Next ›
|
||||
</Link>
|
||||
|
||||
@ -83,12 +83,12 @@ import { computed } from 'vue'
|
||||
import { Link } from '@inertiajs/vue3'
|
||||
|
||||
const props = defineProps({
|
||||
data: Object
|
||||
data: Object,
|
||||
})
|
||||
|
||||
// Core pagination values
|
||||
const links = computed(() => props.data.meta.links || [])
|
||||
const currentIndex = computed(() => links.value.findIndex(link => link.active))
|
||||
const currentIndex = computed(() => links.value.findIndex((link) => link.active))
|
||||
|
||||
const maxVisible = 10
|
||||
const half = Math.floor(maxVisible / 2)
|
||||
@ -115,5 +115,7 @@ const nextPageUrl = computed(() => links.value[currentIndex.value + 1]?.url)
|
||||
const lastPageUrl = computed(() => links.value[links.value.length - 2]?.url) // last item is "Next »", second-last is last numbered
|
||||
|
||||
const isFirstPage = computed(() => links.value[currentIndex.value]?.label === '1')
|
||||
const isLastPage = computed(() => links.value[currentIndex.value]?.label === props.data.meta.last_page)
|
||||
const isLastPage = computed(
|
||||
() => links.value[currentIndex.value]?.label === props.data.meta.last_page
|
||||
)
|
||||
</script>
|
||||
|
@ -34,5 +34,5 @@ export default await Env.create(new URL('../', import.meta.url), {
|
||||
PG_USER: Env.schema.string(),
|
||||
PG_PASSWORD: Env.schema.string(),
|
||||
|
||||
WEBHOOK_URL: Env.schema.string()
|
||||
WEBHOOK_URL: Env.schema.string(),
|
||||
})
|
||||
|
@ -11,6 +11,5 @@ import ReplaysController from '#controllers/replays_controller'
|
||||
import router from '@adonisjs/core/services/router'
|
||||
router.on('/').renderInertia('home')
|
||||
router.get('/replays', [ReplaysController, 'index'])
|
||||
router.get('/list', [ReplaysController, 'list'
|
||||
])
|
||||
router.get('/list', [ReplaysController, 'list'])
|
||||
router.get('/stats', [ReplaysController, 'stats'])
|
Reference in New Issue
Block a user