images = $( sort $( wildcard images/*.jpg) )
output_ocr := $( dir_ocr) /output.txt
tmpfile := $( shell mktemp)
space := $( empty) $( empty)
newline := '\n'
listimgs := $( subst $( space) ,$( newline) , $( images) ) # list of the images, with one filename on each line $(subst $(delimitator),$(replacement),$(list))
OS := $( shell uname)
# Colors: add color to output ie @echo $(color_r) output text
color_w := "\033[0;29m"
color_r := "\033[0;31m"
color_g := "\033[0;32m"
color_b := "\033[0;34m"
# HELP / SELF DOCUMENTATION
# rules where first line contains comment with 2x# (see example in clean rule)
.DEFAULT_GOAL := help # help rule as default when you run: make
.PHONY : help
help :
@grep -E '^[a-zA-Z_-\/]+:.*?## .*$$' $( MAKEFILE_LIST) | sort | awk 'BEGIN {FS = ":.*?## "}; {printf "\033[36m%-30s\033[0m %s\n", $$1, $$2}'
# CLEAN
clean : ## removes output (target) files
rm ocr/output.txt
rm $( wildcard output/*)
rm $( tmpfile)
# ADMINISTRATIVE RECIPES
dirs : ## create the dirs in working dir
@-mkdir -p images/
@-mkdir -p output/
@-mkdir -p ocr/
@echo $( color_r) 'Directories made' : images/ output/
testif :
i f e q ( $( OS ) , D a r w i n )
@echo $( OS)
e n d i f
# POST-PROCESSING RECIPES
ocr/output.txt : ## ocr with tesseract
echo $( listimgs) > $( @D) /list.txt
@echo $( basename $@ .txt)
tesseract $( @D) /list.txt $( basename $@ .txt)
#OUTPUT GENERATION RECIPES
output/tagged-words.txt : ocr /output .txt ## Analyzes OCR'ed text using a Part of Speech (POS) tagger. Outputs a string of tags (e.g. nouns, verbs, adjectives, and adverbs). Dependencies: python3's nltk, nltk's averaged_perceptron_tagger
cat $< | python3 src/wordtagger.py > $( @)
# install nltk's 'averaged_perceptron_tagger':
# $ python 3
# >>> import nltk
# >>> nltk.download('averaged_perceptron_tagger')
output/chatbot.txt : ocr /output .txt ## Comments a text with a simple chatbot. Dependencies: python3's chatterbot
cat $< | python3 src/textbotconversation.py $( @)
output/n7.txt : ocr /output .txt ## DESCRIBE WHAT IT DOES. Dependencies: python3's chatterbot
cat $< | python3 src/n_7.py > $( @)
visualization : $( images ) $( tmpfile ) ##Creates data visualization from images/*.jpg. Dependencies: mplayer
@echo $( tmpfile)
for i in $( images) ; do \
cat $$ i >> $( tmpfile) ; \
done ;
i f e q ( $( OS ) , D a r w i n )
cat $( tmpfile) | mplayer -sws 4 -zoom -vf dsize = 720:720 -demuxer rawvideo -rawvideo w = 56:h= 64:i420:fps= 25 -;
e l s e
cat $( tmpfile) | mplayer -vo x11 -sws 4 -zoom -vf dsize = 720:720 -demuxer rawvideo -rawvideo w = 50:h= 50:i420:fps= 25 -;
e n d i f
ttssr-human-only : ocr /output .txt ## Loop: text to speech-speech recognition. Dependencies: espeak, pocketsphinx
bash src/ttssr-loop-human-only.sh ocr/output.txt