Skip to content

chocolatemelt/mtg-analyst

Repository files navigation

mtg analyst
===========

just playing around with streamed analytics. runs on druid, kafka, and scryfall (as well as related deps).
generate dump with batch.py or individual sets with setparse.py
generate all current commander, core, and expansion sets with genset.py

python packages:
kafka-python
requests

kafka / druid quickstart:
1. download druid, zookeeper, and kafka tarballs and unzip them
2. move zookeeper into druid root as folder zk
3. start druid: ./bin/start-micro-quickstart
4. start kafka: ./bin/kafka-server-start.sh ./config/server.properties
5. start mtg topic: ./bin/kafka-topics.sh --create --zookeeper localhost:2181 --replication-factor 1 --partitions 1 --topic mtg
6. submit mtg supervisor: curl -XPOST -H'Content-Type: application/json' -d @./mtg-spec.json http://localhost:8081/druid/indexer/v1/supervisor
7: stream data as you like

faq:
why tar.bz2 for single files
	i wasn't paying attention, and by the time i did, i didn't really want to change it

these queries suck
	sorry, i'm new to druid and i was never particularly strong at crafting sql

2019-09-24 and 2019-09-25 are the same dataset
	sorry, i overwrote the data of the 24th with the data from the 25th, which makes it somewhat inaccurate as a viable dataset...

About

streamed analytics using scryfall

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published