[07:29:27] o/ [07:38:27] o/ [08:32:51] pfischer: when you have a moment could you take a look at https://gitlab.wikimedia.org/repos/search-platform/cirrus-streaming-updater/-/merge_requests/197? [08:34:50] sure [08:35:55] dcausse: done! [08:38:57] thanks! [10:39:42] lunch [13:18:17] o/ [14:29:19] \o [14:37:04] i forget, how big of runs did we typically do with relforge, like 10k queries? [14:38:33] I guess i was pondering if its worthwhile to loop through the cirrus AST api to be sure we have queries of class "simple_bag_of_words" [14:38:53] o/ [14:39:13] also, separately, the cirrusDumpQueryAST endpoint is returning an ast for the main wiki, and then responses from all the secondary searches (wiktionary, etc) [14:39:25] not a big deal, but probably not intended :) [14:39:43] indeed not really necessary :) [14:39:57] the simple_bag_of_words tag should be in cirrus backend logs already [14:41:33] but for historical data yes... might be tricky to get, cirrus backend logs might not have the same retention as query_clicks [14:42:11] dcausse: i guess i was expecting various inputs, i think in the past we sourced from various places depending on use case (webreq, mjolnir clicks, search sat). But could limit to CSRQ [14:43:15] sure, calling cirrus is probably OK but we'll need to stop early before hitting sister wikis :) [14:44:05] yea probably worth fixing [14:44:56] volans: what's the best way to test https://gerrit.wikimedia.org/r/c/operations/software/spicerack/+/1167299? it'd be nice to be able to run it before merging ideally [14:45:10] or simple heuristics if we don't care too much about dropping good queries (no : " && || AND OR NOT, ~) [14:45:28] yea maybe thats good enough [14:45:39] some regex that rejects queries [14:47:50] yesterday I noticed a bit of randomness in ltr results and was wondering how much of this is caused by the use raw freq features [14:48:40] ryankemper: I'll defer to elukey as I'm currently in another team (I'm not sure how widespread the communication was). But keep in mind that currently spicerack on the bookworm cumin hosts has no elasticsearch support at all, so it can't break anything. That said there are potential ways to test it via spicerack-shell and some manual step. [14:50:05] hmm, i suppose the freq feature do vary by shard, but i suppose i expected them to be close enough [14:52:06] if you refresh you sometimes see some results flipping around, generally not huge movements more like move up or down by 1 pos [14:53:40] yea thats probably unsolvable, too many things vary it must be right near a cutoff point [14:53:51] volans: thanks will follow up with him! [15:05:39] can't remember if mjolnir is pushing a new featureset to the ltr index on every runs (after feature selection) or somehow is able to extract features without storing a predefined featureset [15:13:04] it fetches the featureset from opensearch prior to uploading the model, then uploads a minimized feature set with the model. The features basically get stored inside the model on the plugin side [15:13:29] in part needed because feature selection is run for every model, so it's always different [15:13:38] (although it would be curious to run some analysis over chosen features some day) [15:25:13] ok, so the process to push a new feature set is always "manual" [16:06:41] break, back in ~15-20 [16:25:16] back [17:49:39] lunch, back in ~40 [18:06:05] * ebernhardson realizes there are old notebooks in the relforge repo [18:08:51] and i somehow have 15G of wikidata completion clicks data from 2022 in the my local relforge dir [19:01:29] ebernhardson: 1:1 in https://meet.google.com/stp-swkd-iho ? [19:01:44] gehel: oh, sec [19:01:52] :) [19:02:04] you're not late yet! [20:48:35] huh, never seen that: "parse_exception: Cannot parse 'newcastle united 2002/03': Lexical error at line 1, column 25. Encountered: after : "/03" [21:42:43] relforge reports now generating again, although currently only from a notebook. Will have to ponder how best to run this [21:43:08] also i only wrote the bit that sources queries from mjolnir clicks, but the others shouldn't be that hard. Just need a dataframe with query strings for a single wiki [21:44:07] as for if the relaxed query is better at some things...i dunno :P like `list of nail polish brands` gives `Nail (anatomy)`, a famous manicurist, and a multi-national beauty company as top 3. relaxed gave `Nail polish`, `Nail art` and `Nail salon`. Both sucked :P [21:45:04] but it absolutely does something: Top 3 Sorted Results Differ: 79.0%