-
Notifications
You must be signed in to change notification settings - Fork 0
chocolatemelt/mtg-analyst
Folders and files
| Name | Name | Last commit message | Last commit date | |
|---|---|---|---|---|
Repository files navigation
mtg analyst =========== just playing around with streamed analytics. runs on druid, kafka, and scryfall (as well as related deps). generate dump with batch.py or individual sets with setparse.py generate all current commander, core, and expansion sets with genset.py python packages: kafka-python requests kafka / druid quickstart: 1. download druid, zookeeper, and kafka tarballs and unzip them 2. move zookeeper into druid root as folder zk 3. start druid: ./bin/start-micro-quickstart 4. start kafka: ./bin/kafka-server-start.sh ./config/server.properties 5. start mtg topic: ./bin/kafka-topics.sh --create --zookeeper localhost:2181 --replication-factor 1 --partitions 1 --topic mtg 6. submit mtg supervisor: curl -XPOST -H'Content-Type: application/json' -d @./mtg-spec.json http://localhost:8081/druid/indexer/v1/supervisor 7: stream data as you like faq: why tar.bz2 for single files i wasn't paying attention, and by the time i did, i didn't really want to change it these queries suck sorry, i'm new to druid and i was never particularly strong at crafting sql 2019-09-24 and 2019-09-25 are the same dataset sorry, i overwrote the data of the 24th with the data from the 25th, which makes it somewhat inaccurate as a viable dataset...
About
streamed analytics using scryfall
Resources
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published